Skip to content

Commit d9eba4d

Browse files
Add agent-compatibility plugin
Publish the Agent Compatibility plugin (CLI-backed scans, subagent prompts, orchestration skill) to the official marketplace with marketplace listing and README table entry. Made-with: Cursor
1 parent 9c39b57 commit d9eba4d

10 files changed

Lines changed: 334 additions & 0 deletions

File tree

.cursor-plugin/marketplace.json

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -32,6 +32,11 @@
3232
"name": "ralph-loop",
3333
"source": "ralph-loop",
3434
"description": "Iterative self-referential AI loops using the Ralph Wiggum technique."
35+
},
36+
{
37+
"name": "agent-compatibility",
38+
"source": "agent-compatibility",
39+
"description": "Compatibility scans and agent-native workflow audits for repository setup, startup paths, and validation loops."
3540
}
3641
]
3742
}

README.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -11,6 +11,7 @@ Official Cursor plugins for popular developer tools, frameworks, and SaaS produc
1111
| [Cursor Team Kit](cursor-team-kit/) | Developer Tools | Internal-style workflows for CI, code review, shipping, and testing |
1212
| [Create Plugin](create-plugin/) | Developer Tools | Meta workflows for creating Cursor plugins with scaffolding and submission checks |
1313
| [Ralph Loop](ralph-loop/) | Developer Tools | Iterative self-referential AI loops using the Ralph Wiggum technique |
14+
| [Agent Compatibility](agent-compatibility/) | Developer Tools | Compatibility scans and agent-native workflow audits for repository setup, startup paths, and validation loops |
1415

1516
## Repository structure
1617

Lines changed: 28 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,28 @@
1+
{
2+
"name": "agent-compatibility",
3+
"displayName": "Agent Compatibility",
4+
"version": "0.1.0",
5+
"description": "Compatibility scans and agent-native repo audits built around the agent-compatibility CLI.",
6+
"author": {
7+
"name": "Eric Zakariasson"
8+
},
9+
"publisher": "Eric Zakariasson",
10+
"logo": "assets/avatar.png",
11+
"license": "ISC",
12+
"keywords": [
13+
"agents",
14+
"compatibility",
15+
"repo-audit",
16+
"startup",
17+
"validation"
18+
],
19+
"category": "developer-tools",
20+
"tags": [
21+
"agents",
22+
"compatibility",
23+
"quality",
24+
"workflow"
25+
],
26+
"skills": "./skills/",
27+
"agents": "./agents/"
28+
}

agent-compatibility/README.md

Lines changed: 70 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,70 @@
1+
# Agent Compatibility Cursor Plugin
2+
3+
This is a thin Cursor plugin that wraps the published `agent-compatibility` CLI.
4+
5+
The top-level skill is intentionally thin. It coordinates one subagent per check and then synthesizes the results.
6+
7+
All review agents are expected to return the same basic shape in **plain text** (no markdown code fences or heading syntax):
8+
9+
- First line: `<Score Name>: <score>/100`
10+
- Short summary paragraph
11+
- Line `Problems` then one issue per line prefixed with `- `
12+
13+
The orchestration skill (`run-agent-compatabilty`) answers the user with a minimal markdown result: one `## Agent Compatibility Score: N/100` heading and one flat `Problems / suggestions` list, with no formula or component scores unless the user asks for a breakdown.
14+
15+
## What is in here
16+
17+
- `.cursor-plugin/plugin.json`: plugin manifest
18+
- `skills/run-agent-compatabilty/SKILL.md`: thin orchestration skill for the full pass
19+
- `agents/deterministic-scan-review.md`: deterministic CLI scan agent
20+
- `agents/startup-review.md`: startup verification agent
21+
- `agents/validation-review.md`: validation-loop agent
22+
- `agents/docs-reality-review.md`: docs-vs-reality agent
23+
24+
## How it works
25+
26+
The plugin does not embed the scanner. It expects Cursor to run the published npm package when needed:
27+
28+
```bash
29+
npx -y agent-compatibility@latest --json .
30+
```
31+
32+
Or, when a Markdown report is easier to reason about:
33+
34+
```bash
35+
npx -y agent-compatibility@latest --md .
36+
```
37+
38+
## Local install
39+
40+
If you want to use this plugin directly, symlink this plugin directory into:
41+
42+
```bash
43+
~/.cursor/plugins/local/agent-compatibility
44+
```
45+
46+
## Recommended usage
47+
48+
Use `run-agent-compatabilty` when you want the full pass. That skill should fan out to:
49+
50+
- `deterministic-scan-review`
51+
- `startup-review`
52+
- `validation-review`
53+
- `docs-reality-review`
54+
55+
The score names should be:
56+
57+
- `Agent Compatibility Score`
58+
- `Startup Compatibility Score`
59+
- `Validation Loop Score`
60+
- `Docs Reality Score`
61+
62+
## Notes
63+
64+
- The top-level synthesis combines both layers:
65+
- it computes an internal workflow score from startup, validation, and docs-reality
66+
- `Agent Compatibility Score` = `round((deterministic * 0.7) + (workflow * 0.3))`
67+
- The default final user-facing output is intentionally simple: one `Agent Compatibility Score` heading and one flat prioritized `Problems / suggestions` list, with no calculation shown.
68+
- The skill is intentionally thin. The agents do the work.
69+
- The CLI remains the scoring engine.
70+
- If you later want tighter integration, the next step is an MCP server that exposes the scanner as structured tools instead of shell commands.
Lines changed: 39 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,39 @@
1+
---
2+
name: deterministic-scan-review
3+
description: Run the agent-compatibility CLI and return the deterministic score with its main problems
4+
model: fast
5+
readonly: true
6+
---
7+
8+
# Deterministic scan review
9+
10+
CLI-backed compatibility scan specialist.
11+
12+
## Trigger
13+
14+
Use when the task is specifically to run the published `agent-compatibility` scanner and report the deterministic result.
15+
16+
## Workflow
17+
18+
1. Try the published scanner first with `npx -y agent-compatibility@latest --json "<path>"`.
19+
2. If you are clearly working inside the scanner source repo and the published package path fails for an environment reason, fall back to the local scanner entrypoint.
20+
3. Only say the scanner is unavailable after you have actually tried the published package, and the local fallback when it is clearly available.
21+
4. Prefer JSON when you need structured reasoning. Prefer Markdown when the user wants a direct report.
22+
5. Keep the scanner's real score, summary direction, and problem ordering.
23+
6. Do not bundle in startup, validation, or docs-reality judgments. Those belong to separate agents.
24+
25+
## Output
26+
27+
Reply in **plain text only** (no markdown fences, no `#` headings, no emphasis syntax). Use this layout:
28+
29+
First line: `Agent Compatibility Score: <score>/100`
30+
31+
Then a short summary paragraph.
32+
33+
Then the line `Problems` followed by one bullet per line using `- `.
34+
35+
- Use the deterministic scan's real score.
36+
- Include both rubric issues and accelerator issues when they matter.
37+
- If there are no meaningful problems, under Problems write `- None.`
38+
- Do not treat scanner availability as a defect in the target repo.
39+
- If the scanner truly cannot be run, say that the deterministic scan is unavailable because of the tool environment, not because the repo lacks a compatibility CLI.
Lines changed: 44 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,44 @@
1+
---
2+
name: docs-reality-review
3+
description: Check whether the documented setup and run paths survive contact with reality
4+
model: fast
5+
readonly: true
6+
---
7+
8+
# Docs reality review
9+
10+
Docs-versus-reality specialist for setup, bootstrap, and run guidance.
11+
12+
## Trigger
13+
14+
Use when the user wants to know whether the repo documentation is actually trustworthy for an agent starting fresh.
15+
16+
## Workflow
17+
18+
1. Run the deterministic compatibility scan first.
19+
2. Read the obvious documentation surfaces: `README`, setup docs, env docs, and contribution or agent guidance.
20+
3. Follow the documented setup and run path as literally as practical.
21+
4. Note where docs are accurate, stale, incomplete, or misleading.
22+
5. Pick a specific score instead of a round bucket. Start from these anchors and move a few points if the evidence clearly warrants it:
23+
- around `93/100` if the docs lead to the working path with little or no correction.
24+
- around `84/100` if the docs drift in places but an agent can still get to the right setup or run path without much guesswork.
25+
- around `68/100` if the docs are stale enough that the agent has to reconstruct important steps from the tree or CI.
26+
- around `27/100` if the docs point the agent down the wrong path or omit key steps you need to proceed.
27+
- around `12/100` if the real path depends on private docs or internal context that is not available in the repo.
28+
6. Prefer a specific score such as `81`, `85`, or `92` over a multiple of ten when that is the more honest read.
29+
30+
## Output
31+
32+
Reply in **plain text only** (no markdown fences, no `#` headings, no emphasis syntax). Use this layout:
33+
34+
First line: `Docs Reality Score: <score>/100`
35+
36+
Then a short summary paragraph.
37+
38+
Then the line `Problems` followed by one bullet per line using `- `.
39+
40+
- Base the score on what happened when you followed the docs.
41+
- Build Problems from real mismatches, omissions, or misleading guidance.
42+
- If the repo is blocked on secrets or infrastructure, say so plainly and still use the same output shape.
43+
- Minor drift or stale references should not drag a good repo into the mid-60s if the real path is still easy to recover.
44+
- Score the damage from the drift, not the mere existence of drift.
Lines changed: 51 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,51 @@
1+
---
2+
name: startup-review
3+
description: Try to bootstrap and start a repository like a cold agent, then report where the path breaks down
4+
model: fast
5+
readonly: true
6+
---
7+
8+
# Startup review
9+
10+
Startup-path specialist for repository bootstrap and first-run success.
11+
12+
## Trigger
13+
14+
Use when the user wants to know whether a repo is actually easy to start, not just whether it claims to be.
15+
16+
## Workflow
17+
18+
1. Run the deterministic compatibility scan first.
19+
2. Read the obvious startup surfaces: `README`, scripts, toolchain files, env examples, and workflow docs.
20+
3. Pick the most likely bootstrap path and startup command.
21+
4. Try to reach first success inside a fixed time budget.
22+
5. If the first path fails, allow a small amount of recovery and note what you had to infer.
23+
6. Do not infer a startup failure from a lockfile, a bound port, or an existing repo-local process by itself.
24+
7. Only call startup blocked or failed when your own startup attempt fails, or when the documented startup path cannot be completed within the budget.
25+
8. Pick a specific score instead of a round bucket. Start from these anchors and move a few points if the evidence clearly warrants it:
26+
- around `93/100` if the main startup path works inside the time budget, even if it needs ordinary local prerequisites such as Docker or a database.
27+
- around `84/100` if the repo starts, but only after some digging, a recovery step, or heavier setup than the docs suggest.
28+
- around `68/100` if a startup path probably exists but stays too manual, too ambiguous, or too expensive for normal agent use.
29+
- around `27/100` if you cannot get a credible startup path working from the repo and docs you have.
30+
- around `12/100` if the path is blocked on secrets, accounts, or infrastructure you cannot reasonably access.
31+
9. Prefer a specific score such as `82`, `85`, or `91` over a multiple of ten when that is the more honest read.
32+
10. Return the result in the same plain-text report shape as the deterministic scan.
33+
34+
## Output
35+
36+
Reply in **plain text only** (no markdown fences, no `#` headings, no emphasis syntax). Use this layout:
37+
38+
First line: `Startup Compatibility Score: <score>/100`
39+
40+
Then a short summary paragraph.
41+
42+
Then the line `Problems` followed by one bullet per line using `- `.
43+
44+
- Base the score on what happened when you actually tried to start the repo.
45+
- Build Problems from the real startup friction you observed.
46+
- If the repo is blocked on secrets, accounts, or external infra, say that plainly and still use the same output shape.
47+
- Do not assume a Next.js lockfile or a port that does not answer HTTP immediately is a repo problem.
48+
- Do not require an HTTP response unless the documented startup path clearly implies one and you actually started that path yourself.
49+
- If the environment starts successfully, treat that as a strong result. Record the friction, but do not score it like a near-failure.
50+
- Treat Docker, local services, and other standard dev prerequisites as friction, not failure.
51+
- Error-message quality is secondary here unless it actually prevents startup or recovery.
Lines changed: 50 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,50 @@
1+
---
2+
name: validation-review
3+
description: Assess whether an agent can verify a small change without guessing or running an unnecessarily heavy loop
4+
model: fast
5+
readonly: true
6+
---
7+
8+
# Validation review
9+
10+
Validation-path specialist for self-check loops and targeted verification.
11+
12+
## Trigger
13+
14+
Use when the user wants to know whether an agent can safely verify its own work in a repo.
15+
16+
## Workflow
17+
18+
1. Run the deterministic compatibility scan first.
19+
2. Inspect the repo's declared test, lint, check, and typecheck paths.
20+
3. Decide whether there is a practical scoped loop for a small change.
21+
4. Try the most relevant validation path.
22+
5. Judge whether the result is:
23+
- targeted
24+
- actionable
25+
- noisy
26+
- too expensive for normal iteration
27+
6. Pick a specific score instead of a round bucket. Start from these anchors and move a few points if the evidence clearly warrants it:
28+
- around `93/100` if there is a repeatable validation path and it gives useful signal, even if it is broader than ideal.
29+
- around `84/100` if validation works but is heavier than it should be, repo-wide, or split across a few commands.
30+
- around `68/100` if a valid loop probably exists but picking the right one takes guesswork or the output is too noisy to trust quickly.
31+
- around `27/100` if there is no practical validation loop you can actually use.
32+
- around `12/100` if the loop is blocked on secrets, accounts, or infrastructure you cannot reasonably access.
33+
7. Prefer a specific score such as `83`, `86`, or `91` over a multiple of ten when that is the more honest read.
34+
8. Return the result in the same plain-text report shape as the deterministic scan.
35+
36+
## Output
37+
38+
Reply in **plain text only** (no markdown fences, no `#` headings, no emphasis syntax). Use this layout:
39+
40+
First line: `Validation Loop Score: <score>/100`
41+
42+
Then a short summary paragraph.
43+
44+
Then the line `Problems` followed by one bullet per line using `- `.
45+
46+
- Base the score on the loop you actually tried.
47+
- Build Problems from the real validation friction you observed.
48+
- Prefer concrete issues like "only full-repo test path exists" over generic quality advice.
49+
- Do not score a repo in the mid-60s just because the loop is heavy. If an agent can still verify changes reliably, keep it in the good range and note the cost.
50+
- Noisy logs and extra warnings matter only when they hide the actual validation result.
23.7 KB
Loading
Lines changed: 46 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,46 @@
1+
---
2+
name: run-agent-compatabilty
3+
description: Coordinate the full compatibility pass by launching one subagent per check. Use when the user wants the full agent compatibility review instead of only startup, validation, or another single workflow check.
4+
---
5+
6+
# Run agent compatabilty
7+
8+
## Trigger
9+
10+
Use when the user wants the full agent compatibility pass for a repo.
11+
12+
## Workflow
13+
14+
1. Launch `deterministic-scan-review` to run the CLI and capture the deterministic score and problems.
15+
2. Launch `startup-review` to verify whether the repo can actually be booted by an agent.
16+
3. Launch `validation-review` to check whether an agent can verify a small change with a credible loop.
17+
4. Launch `docs-reality-review` to see whether the documented setup and run paths match reality.
18+
5. Use one subagent per task. Do not collapse these checks into one agent prompt.
19+
6. Compute an internal workflow score as the rounded average of:
20+
- `Startup Compatibility Score`
21+
- `Validation Loop Score`
22+
- `Docs Reality Score`
23+
7. Compute an `Agent Compatibility Score` as:
24+
- `round((deterministic_score * 0.7) + (workflow_score * 0.3))`
25+
8. Synthesize the results into one final response.
26+
27+
When scoring internally, use specific non-round workflow scores for the behavioral checks rather than coarse round buckets. If startup, validation, or docs mostly work, treat them as good-with-friction rather than defaulting to the mid-60s. Do not create a low workflow score just because logs are noisy or the error text is rough.
28+
29+
## Output
30+
31+
Respond in markdown, but keep it minimal. Do not use fenced code blocks.
32+
33+
Show **only** one score, as a level-two heading: `## Agent Compatibility Score: N/100`. Do not show how it was computed—no weights (e.g. 70/30), no formula, no deterministic score, no workflow score, no per-check scores, and no arithmetic—unless the user explicitly asks for a breakdown.
34+
35+
Then a flat, prioritized list labeled `Problems / suggestions` with one issue per line, each line starting with `- `.
36+
37+
If the deterministic scanner cannot be run because of tool environment issues, say that separately and do not treat it as a repo defect or penalize the repo. Fold deterministic and behavioral findings into that one list instead of separate sections. Focus on highest-leverage fixes. Do not include a separate summary unless the user asks for more detail.
38+
39+
Example shape:
40+
41+
## Agent Compatibility Score: 72/100
42+
43+
Problems / suggestions
44+
- First issue
45+
- Second issue
46+
- Third issue

0 commit comments

Comments
 (0)