skill-comply▌
affaan-m/everything-claude-code · updated Apr 8, 2026
Measures whether coding agents actually follow skills, rules, or agent definitions by:
skill-comply: Automated Compliance Measurement
Measures whether coding agents actually follow skills, rules, or agent definitions by:
- Auto-generating expected behavioral sequences (specs) from any .md file
- Auto-generating scenarios with decreasing prompt strictness (supportive → neutral → competing)
- Running
claude -pand capturing tool call traces via stream-json - Classifying tool calls against spec steps using LLM (not regex)
- Checking temporal ordering deterministically
- Generating self-contained reports with spec, prompts, and timelines
Supported Targets
- Skills (
skills/*/SKILL.md): Workflow skills like search-first, TDD guides - Rules (
rules/common/*.md): Mandatory rules like testing.md, security.md, git-workflow.md - Agent definitions (
agents/*.md): Whether an agent gets invoked when expected (internal workflow verification not yet supported)
When to Activate
- User runs
/skill-comply <path> - User asks "is this rule actually being followed?"
- After adding new rules/skills, to verify agent compliance
- Periodically as part of quality maintenance
Usage
# Full run
uv run python -m scripts.run ~/.claude/rules/common/testing.md
# Dry run (no cost, spec + scenarios only)
uv run python -m scripts.run --dry-run ~/.claude/skills/search-first/SKILL.md
# Custom models
uv run python -m scripts.run --gen-model haiku --model sonnet <path>
Key Concept: Prompt Independence
Measures whether a skill/rule is followed even when the prompt doesn't explicitly support it.
Report Contents
Reports are self-contained and include:
- Expected behavioral sequence (auto-generated spec)
- Scenario prompts (what was asked at each strictness level)
- Compliance scores per scenario
- Tool call timelines with LLM classification labels
Advanced (optional)
For users familiar with hooks, reports also include hook promotion recommendations for steps with low compliance. This is informational — the main value is the compliance visibility itself.
Discussion
Product Hunt–style comments (not star reviews)- No comments yet — start the thread.
Ratings
4.5★★★★★36 reviews- ★★★★★Dhruvi Jain· Dec 20, 2024
I recommend skill-comply for anyone iterating fast on agent tooling; clear intent and a small, reviewable surface area.
- ★★★★★Anaya Okafor· Dec 8, 2024
Solid pick for teams standardizing on skills: skill-comply is focused, and the summary matches what you get after install.
- ★★★★★Arjun Ndlovu· Dec 4, 2024
Registry listing for skill-comply matched our evaluation — installs cleanly and behaves as described in the markdown.
- ★★★★★Anaya Ndlovu· Nov 27, 2024
skill-comply has been reliable in day-to-day use. Documentation quality is above average for community skills.
- ★★★★★Rahul Santra· Nov 23, 2024
Keeps context tight: skill-comply is the kind of skill you can hand to a new teammate without a long onboarding doc.
- ★★★★★Diego Abbas· Nov 23, 2024
skill-comply reduced setup friction for our internal harness; good balance of opinion and flexibility.
- ★★★★★Oshnikdeep· Nov 11, 2024
skill-comply fits our agent workflows well — practical, well scoped, and easy to wire into existing repos.
- ★★★★★Kwame Thomas· Oct 18, 2024
skill-comply fits our agent workflows well — practical, well scoped, and easy to wire into existing repos.
- ★★★★★Pratham Ware· Oct 14, 2024
We added skill-comply from the explainx registry; install was straightforward and the SKILL.md answered most questions upfront.
- ★★★★★Camila Sharma· Oct 14, 2024
skill-comply is among the better-maintained entries we tried; worth keeping pinned for repeat workflows.
showing 1-10 of 36