Compare commits
236 Commits
4fb7a53c55
...
marketplac
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
51f906c9ff | ||
|
|
78971c9027 | ||
|
|
f508a3f759 | ||
|
|
f47f829d81 | ||
|
|
74fb71731a | ||
|
|
6a3d5b4bf3 | ||
|
|
5c67d287c4 | ||
|
|
04f00e7632 | ||
|
|
a83e11e982 | ||
|
|
72d4b0dfd2 | ||
|
|
7a81cd1aba | ||
|
|
ac53635737 | ||
|
|
ad53d3d657 | ||
|
|
b2b23ddbd3 | ||
|
|
80818617bc | ||
|
|
6e060e9f9e | ||
|
|
df4c466b42 | ||
|
|
859ef601b2 | ||
|
|
38a47b11ca | ||
|
|
bbdefbf8b9 | ||
|
|
6462de20a6 | ||
|
|
db61ad3655 | ||
|
|
5bc3a0f469 | ||
|
|
e3b6f19412 | ||
|
|
aa71dbc24f | ||
|
|
0fc6717542 | ||
|
|
3023bfc8c1 | ||
|
|
b762c7647c | ||
|
|
6d38bc7b59 | ||
|
|
f6cca58820 | ||
|
|
bf6d7d5253 | ||
|
|
e4ee77aa1e | ||
|
|
b290690655 | ||
|
|
350465e81a | ||
|
|
6f561f94b4 | ||
|
|
82c1fe86df | ||
|
|
8c9f905859 | ||
|
|
dfff20e1ad | ||
|
|
ff99b0a2e3 | ||
|
|
fdbd584bac | ||
|
|
30c06e5122 | ||
|
|
164a1d651a | ||
|
|
108d872075 | ||
|
|
61ab6e9bab | ||
|
|
4ecc2008ab | ||
|
|
ebb109f3a4 | ||
|
|
6aed93123e | ||
|
|
84ca52efdb | ||
|
|
637653d2ed | ||
|
|
c2582fab67 | ||
|
|
c77e01bb61 | ||
|
|
462456f582 | ||
|
|
b7e43910fb | ||
|
|
a47f7d67a2 | ||
|
|
0c333b08c9 | ||
|
|
6969014532 | ||
|
|
e3e7748c56 | ||
|
|
d4c12c39fd | ||
|
|
db8c84acb4 | ||
|
|
42013612bd | ||
|
|
c271bd4729 | ||
|
|
2ae6fc4458 | ||
|
|
d3aff58d9e | ||
|
|
684814d951 | ||
|
|
699f484033 | ||
|
|
8f4818c6e2 | ||
|
|
95ad09d3e7 | ||
|
|
187571ce97 | ||
|
|
0dff9431ce | ||
|
|
bd3088a851 | ||
|
|
fca3a4019c | ||
|
|
ec8d68580f | ||
|
|
d2c4cee6f9 | ||
|
|
01002450cd | ||
|
|
269fa55436 | ||
|
|
61a7443076 | ||
|
|
24860ec3f1 | ||
|
|
7c04c3158f | ||
|
|
c8261b63d9 | ||
|
|
d0b302ad88 | ||
|
|
4d80a59e51 | ||
|
|
3d0f190097 | ||
|
|
0c0c69be37 | ||
|
|
52e3e14a07 | ||
|
|
bc7b1dbd37 | ||
|
|
9150a1ea54 | ||
|
|
edc3990d55 | ||
|
|
026602e624 | ||
|
|
4f292ee4f7 | ||
|
|
39935b8fd8 | ||
|
|
1886c747d0 | ||
|
|
391adf96e1 | ||
|
|
970816deb9 | ||
|
|
6d03e255aa | ||
|
|
a8f49093de | ||
|
|
0fd38a77c9 | ||
|
|
456ee8aab3 | ||
|
|
30852b7293 | ||
|
|
0ab91847f2 | ||
|
|
a84682cd35 | ||
|
|
1e829ba4fe | ||
|
|
3e99c11c0b | ||
|
|
e426799f85 | ||
|
|
392d682e10 | ||
|
|
4b4d1ae270 | ||
|
|
42cc74c7c7 | ||
|
|
b119d00f2a | ||
|
|
d2ab6c0768 | ||
|
|
55be9a2caf | ||
|
|
e1e5eacd33 | ||
|
|
c21fbe54e9 | ||
|
|
7661a399de | ||
|
|
7dc59fc12e | ||
|
|
8c3f2a584e | ||
|
|
a95a162428 | ||
|
|
4107781855 | ||
|
|
309ea28f74 | ||
|
|
5a982e2925 | ||
|
|
16b1b5a9cc | ||
|
|
3e8a9b7cfa | ||
|
|
86d8518282 | ||
|
|
5884457ffc | ||
|
|
598222e11c | ||
|
|
1db76800f9 | ||
|
|
e94ca04096 | ||
|
|
d96671b9e9 | ||
|
|
1f50483030 | ||
|
|
4fc6ddc5db | ||
|
|
c4b1358431 | ||
|
|
7266062868 | ||
|
|
4fc70939eb | ||
|
|
50cb89efde | ||
|
|
94aedd5a7b | ||
|
|
c649cfc17f | ||
|
|
b07f43ddf5 | ||
|
|
69f2a96e66 | ||
|
|
ca57c67c1c | ||
|
|
4e9899f346 | ||
|
|
020eb8836e | ||
|
|
2fd29ff6ed | ||
|
|
97f9ab34f7 | ||
|
|
eab77bc5b5 | ||
|
|
a3fb1bbfd6 | ||
|
|
7c58eeeec6 | ||
|
|
b703976b6d | ||
|
|
ea21196cf1 | ||
|
|
8fd1670ee2 | ||
|
|
f5e2f257eb | ||
|
|
168c946033 | ||
|
|
1a0ddb9de1 | ||
|
|
9a16de4627 | ||
|
|
87c19f99e6 | ||
|
|
c8ebc5b2d5 | ||
|
|
98e0a169cb | ||
|
|
465bd3d6bb | ||
|
|
a2c4978ace | ||
|
|
0e32da268e | ||
|
|
5f7428269d | ||
|
|
a84288cf8b | ||
|
|
1514e51a94 | ||
|
|
5846cde548 | ||
|
|
718cbfc73d | ||
|
|
01f15fdca9 | ||
|
|
64c9b326a8 | ||
|
|
6fdffab0a9 | ||
|
|
de3d4a99a3 | ||
|
|
e06e5b8c61 | ||
|
|
15b2296bd8 | ||
|
|
30837ef2e9 | ||
|
|
2e2a58beab | ||
|
|
15769e2ece | ||
|
|
62a66c8f7f | ||
|
|
b817b9e3d9 | ||
|
|
30d33270b9 | ||
|
|
6b46fb2ccf | ||
|
|
2370da9c3c | ||
|
|
1ea9806fb6 | ||
|
|
19f581b7af | ||
|
|
c55be29de0 | ||
|
|
f442409d9f | ||
|
|
803b1d662e | ||
|
|
d40b39fa19 | ||
|
|
305fea486f | ||
|
|
4b60bcaf6c | ||
|
|
e2abb3df91 | ||
|
|
e1d5bdedb3 | ||
|
|
8a530f7e25 | ||
|
|
c597091849 | ||
|
|
1107e3dd31 | ||
|
|
e081e32a30 | ||
|
|
6fe51a0602 | ||
|
|
a3701e220d | ||
|
|
9196ed8ad8 | ||
|
|
03f6ec64b3 | ||
|
|
8f5dd37274 | ||
|
|
3e384309d6 | ||
|
|
004824b242 | ||
|
|
83a65fe9d4 | ||
|
|
3b4e0ae11f | ||
|
|
63e76cf67f | ||
|
|
1aed2353e2 | ||
|
|
005b5ce2a1 | ||
|
|
e900853604 | ||
|
|
2f05f215b3 | ||
|
|
4db39f8b42 | ||
|
|
06d4aea70c | ||
|
|
27319bd85f | ||
|
|
3914dfdebe | ||
|
|
5abddbcbd9 | ||
|
|
f0b6ce9689 | ||
|
|
da94da90db | ||
|
|
d83c1a29c3 | ||
|
|
174cd4cff4 | ||
|
|
d53ef1e837 | ||
|
|
c2c211107f | ||
|
|
e84075660a | ||
|
|
0099e2559a | ||
|
|
a77eacb856 | ||
|
|
7a41f64f06 | ||
|
|
ee76195daf | ||
|
|
d314d7fa2a | ||
|
|
e1906592cb | ||
|
|
83277feee8 | ||
|
|
4f5efbf852 | ||
|
|
a7b76a059c | ||
|
|
7549f38623 | ||
|
|
dbb25c63dd | ||
|
|
7055df5d8e | ||
|
|
5d984ab2da | ||
|
|
a21bc5844d | ||
|
|
f859619a40 | ||
|
|
877e265ec1 | ||
|
|
bc655f714e | ||
|
|
e4d730d5b4 | ||
|
|
4f7c598f27 | ||
|
|
e469ab0cc0 |
@@ -1,5 +1,5 @@
|
||||
{
|
||||
"name": "every-marketplace",
|
||||
"name": "compound-engineering-plugin",
|
||||
"owner": {
|
||||
"name": "Kieran Klaassen",
|
||||
"url": "https://github.com/kieranklaassen"
|
||||
@@ -11,8 +11,8 @@
|
||||
"plugins": [
|
||||
{
|
||||
"name": "compound-engineering",
|
||||
"description": "AI-powered development tools that get smarter with every use. Make each unit of engineering work easier than the last. Includes 29 specialized agents, 22 commands, and 19 skills.",
|
||||
"version": "2.34.0",
|
||||
"description": "AI-powered development tools that get smarter with every use. Make each unit of engineering work easier than the last. Includes 29 specialized agents and 44 skills.",
|
||||
"version": "2.42.0",
|
||||
"author": {
|
||||
"name": "Kieran Klaassen",
|
||||
"url": "https://github.com/kieranklaassen",
|
||||
|
||||
@@ -1,211 +0,0 @@
|
||||
---
|
||||
name: release-docs
|
||||
description: Build and update the documentation site with current plugin components
|
||||
argument-hint: "[optional: --dry-run to preview changes without writing]"
|
||||
---
|
||||
|
||||
# Release Documentation Command
|
||||
|
||||
You are a documentation generator for the compound-engineering plugin. Your job is to ensure the documentation site at `plugins/compound-engineering/docs/` is always up-to-date with the actual plugin components.
|
||||
|
||||
## Overview
|
||||
|
||||
The documentation site is a static HTML/CSS/JS site based on the Evil Martians LaunchKit template. It needs to be regenerated whenever:
|
||||
|
||||
- Agents are added, removed, or modified
|
||||
- Commands are added, removed, or modified
|
||||
- Skills are added, removed, or modified
|
||||
- MCP servers are added, removed, or modified
|
||||
|
||||
## Step 1: Inventory Current Components
|
||||
|
||||
First, count and list all current components:
|
||||
|
||||
```bash
|
||||
# Count agents
|
||||
ls plugins/compound-engineering/agents/*.md | wc -l
|
||||
|
||||
# Count commands
|
||||
ls plugins/compound-engineering/commands/*.md | wc -l
|
||||
|
||||
# Count skills
|
||||
ls -d plugins/compound-engineering/skills/*/ 2>/dev/null | wc -l
|
||||
|
||||
# Count MCP servers
|
||||
ls -d plugins/compound-engineering/mcp-servers/*/ 2>/dev/null | wc -l
|
||||
```
|
||||
|
||||
Read all component files to get their metadata:
|
||||
|
||||
### Agents
|
||||
For each agent file in `plugins/compound-engineering/agents/*.md`:
|
||||
- Extract the frontmatter (name, description)
|
||||
- Note the category (Review, Research, Workflow, Design, Docs)
|
||||
- Get key responsibilities from the content
|
||||
|
||||
### Commands
|
||||
For each command file in `plugins/compound-engineering/commands/*.md`:
|
||||
- Extract the frontmatter (name, description, argument-hint)
|
||||
- Categorize as Workflow or Utility command
|
||||
|
||||
### Skills
|
||||
For each skill directory in `plugins/compound-engineering/skills/*/`:
|
||||
- Read the SKILL.md file for frontmatter (name, description)
|
||||
- Note any scripts or supporting files
|
||||
|
||||
### MCP Servers
|
||||
For each MCP server in `plugins/compound-engineering/mcp-servers/*/`:
|
||||
- Read the configuration and README
|
||||
- List the tools provided
|
||||
|
||||
## Step 2: Update Documentation Pages
|
||||
|
||||
### 2a. Update `docs/index.html`
|
||||
|
||||
Update the stats section with accurate counts:
|
||||
```html
|
||||
<div class="stats-grid">
|
||||
<div class="stat-card">
|
||||
<span class="stat-number">[AGENT_COUNT]</span>
|
||||
<span class="stat-label">Specialized Agents</span>
|
||||
</div>
|
||||
<!-- Update all stat cards -->
|
||||
</div>
|
||||
```
|
||||
|
||||
Ensure the component summary sections list key components accurately.
|
||||
|
||||
### 2b. Update `docs/pages/agents.html`
|
||||
|
||||
Regenerate the complete agents reference page:
|
||||
- Group agents by category (Review, Research, Workflow, Design, Docs)
|
||||
- Include for each agent:
|
||||
- Name and description
|
||||
- Key responsibilities (bullet list)
|
||||
- Usage example: `claude agent [agent-name] "your message"`
|
||||
- Use cases
|
||||
|
||||
### 2c. Update `docs/pages/commands.html`
|
||||
|
||||
Regenerate the complete commands reference page:
|
||||
- Group commands by type (Workflow, Utility)
|
||||
- Include for each command:
|
||||
- Name and description
|
||||
- Arguments (if any)
|
||||
- Process/workflow steps
|
||||
- Example usage
|
||||
|
||||
### 2d. Update `docs/pages/skills.html`
|
||||
|
||||
Regenerate the complete skills reference page:
|
||||
- Group skills by category (Development Tools, Content & Workflow, Image Generation)
|
||||
- Include for each skill:
|
||||
- Name and description
|
||||
- Usage: `claude skill [skill-name]`
|
||||
- Features and capabilities
|
||||
|
||||
### 2e. Update `docs/pages/mcp-servers.html`
|
||||
|
||||
Regenerate the MCP servers reference page:
|
||||
- For each server:
|
||||
- Name and purpose
|
||||
- Tools provided
|
||||
- Configuration details
|
||||
- Supported frameworks/services
|
||||
|
||||
## Step 3: Update Metadata Files
|
||||
|
||||
Ensure counts are consistent across:
|
||||
|
||||
1. **`plugins/compound-engineering/.claude-plugin/plugin.json`**
|
||||
- Update `description` with correct counts
|
||||
- Update `components` object with counts
|
||||
- Update `agents`, `commands` arrays with current items
|
||||
|
||||
2. **`.claude-plugin/marketplace.json`**
|
||||
- Update plugin `description` with correct counts
|
||||
|
||||
3. **`plugins/compound-engineering/README.md`**
|
||||
- Update intro paragraph with counts
|
||||
- Update component lists
|
||||
|
||||
## Step 4: Validate
|
||||
|
||||
Run validation checks:
|
||||
|
||||
```bash
|
||||
# Validate JSON files
|
||||
cat .claude-plugin/marketplace.json | jq .
|
||||
cat plugins/compound-engineering/.claude-plugin/plugin.json | jq .
|
||||
|
||||
# Verify counts match
|
||||
echo "Agents in files: $(ls plugins/compound-engineering/agents/*.md | wc -l)"
|
||||
grep -o "[0-9]* specialized agents" plugins/compound-engineering/docs/index.html
|
||||
|
||||
echo "Commands in files: $(ls plugins/compound-engineering/commands/*.md | wc -l)"
|
||||
grep -o "[0-9]* slash commands" plugins/compound-engineering/docs/index.html
|
||||
```
|
||||
|
||||
## Step 5: Report Changes
|
||||
|
||||
Provide a summary of what was updated:
|
||||
|
||||
```
|
||||
## Documentation Release Summary
|
||||
|
||||
### Component Counts
|
||||
- Agents: X (previously Y)
|
||||
- Commands: X (previously Y)
|
||||
- Skills: X (previously Y)
|
||||
- MCP Servers: X (previously Y)
|
||||
|
||||
### Files Updated
|
||||
- docs/index.html - Updated stats and component summaries
|
||||
- docs/pages/agents.html - Regenerated with X agents
|
||||
- docs/pages/commands.html - Regenerated with X commands
|
||||
- docs/pages/skills.html - Regenerated with X skills
|
||||
- docs/pages/mcp-servers.html - Regenerated with X servers
|
||||
- plugin.json - Updated counts and component lists
|
||||
- marketplace.json - Updated description
|
||||
- README.md - Updated component lists
|
||||
|
||||
### New Components Added
|
||||
- [List any new agents/commands/skills]
|
||||
|
||||
### Components Removed
|
||||
- [List any removed agents/commands/skills]
|
||||
```
|
||||
|
||||
## Dry Run Mode
|
||||
|
||||
If `--dry-run` is specified:
|
||||
- Perform all inventory and validation steps
|
||||
- Report what WOULD be updated
|
||||
- Do NOT write any files
|
||||
- Show diff previews of proposed changes
|
||||
|
||||
## Error Handling
|
||||
|
||||
- If component files have invalid frontmatter, report the error and skip
|
||||
- If JSON validation fails, report and abort
|
||||
- Always maintain a valid state - don't partially update
|
||||
|
||||
## Post-Release
|
||||
|
||||
After successful release:
|
||||
1. Suggest updating CHANGELOG.md with documentation changes
|
||||
2. Remind to commit with message: `docs: Update documentation site to match plugin components`
|
||||
3. Remind to push changes
|
||||
|
||||
## Usage Examples
|
||||
|
||||
```bash
|
||||
# Full documentation release
|
||||
claude /release-docs
|
||||
|
||||
# Preview changes without writing
|
||||
claude /release-docs --dry-run
|
||||
|
||||
# After adding new agents
|
||||
claude /release-docs
|
||||
```
|
||||
25
.cursor-plugin/marketplace.json
Normal file
25
.cursor-plugin/marketplace.json
Normal file
@@ -0,0 +1,25 @@
|
||||
{
|
||||
"name": "compound-engineering",
|
||||
"owner": {
|
||||
"name": "Kieran Klaassen",
|
||||
"email": "kieran@every.to",
|
||||
"url": "https://github.com/kieranklaassen"
|
||||
},
|
||||
"metadata": {
|
||||
"description": "Cursor plugin marketplace for Every Inc plugins",
|
||||
"version": "1.0.0",
|
||||
"pluginRoot": "plugins"
|
||||
},
|
||||
"plugins": [
|
||||
{
|
||||
"name": "compound-engineering",
|
||||
"source": "compound-engineering",
|
||||
"description": "AI-powered development tools that get smarter with every use. Includes specialized agents, commands, skills, and Context7 MCP."
|
||||
},
|
||||
{
|
||||
"name": "coding-tutor",
|
||||
"source": "coding-tutor",
|
||||
"description": "Personalized coding tutorials with spaced repetition quizzes using your real codebase."
|
||||
}
|
||||
]
|
||||
}
|
||||
6
.github/.release-please-manifest.json
vendored
Normal file
6
.github/.release-please-manifest.json
vendored
Normal file
@@ -0,0 +1,6 @@
|
||||
{
|
||||
".": "2.42.0",
|
||||
"plugins/compound-engineering": "2.42.0",
|
||||
"plugins/coding-tutor": "1.2.1",
|
||||
".claude-plugin": "1.0.0"
|
||||
}
|
||||
64
.github/release-please-config.json
vendored
Normal file
64
.github/release-please-config.json
vendored
Normal file
@@ -0,0 +1,64 @@
|
||||
{
|
||||
"$schema": "https://raw.githubusercontent.com/googleapis/release-please/main/schemas/config.json",
|
||||
"include-component-in-tag": true,
|
||||
"packages": {
|
||||
".": {
|
||||
"release-type": "simple",
|
||||
"package-name": "cli",
|
||||
"skip-changelog": true,
|
||||
"extra-files": [
|
||||
{
|
||||
"type": "json",
|
||||
"path": "package.json",
|
||||
"jsonpath": "$.version"
|
||||
}
|
||||
]
|
||||
},
|
||||
"plugins/compound-engineering": {
|
||||
"release-type": "simple",
|
||||
"package-name": "compound-engineering",
|
||||
"skip-changelog": true,
|
||||
"extra-files": [
|
||||
{
|
||||
"type": "json",
|
||||
"path": ".claude-plugin/plugin.json",
|
||||
"jsonpath": "$.version"
|
||||
},
|
||||
{
|
||||
"type": "json",
|
||||
"path": ".cursor-plugin/plugin.json",
|
||||
"jsonpath": "$.version"
|
||||
}
|
||||
]
|
||||
},
|
||||
"plugins/coding-tutor": {
|
||||
"release-type": "simple",
|
||||
"package-name": "coding-tutor",
|
||||
"skip-changelog": true,
|
||||
"extra-files": [
|
||||
{
|
||||
"type": "json",
|
||||
"path": ".claude-plugin/plugin.json",
|
||||
"jsonpath": "$.version"
|
||||
},
|
||||
{
|
||||
"type": "json",
|
||||
"path": ".cursor-plugin/plugin.json",
|
||||
"jsonpath": "$.version"
|
||||
}
|
||||
]
|
||||
},
|
||||
".claude-plugin": {
|
||||
"release-type": "simple",
|
||||
"package-name": "marketplace",
|
||||
"skip-changelog": true,
|
||||
"extra-files": [
|
||||
{
|
||||
"type": "json",
|
||||
"path": "marketplace.json",
|
||||
"jsonpath": "$.metadata.version"
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
||||
}
|
||||
28
.github/workflows/ci.yml
vendored
28
.github/workflows/ci.yml
vendored
@@ -7,6 +7,31 @@ on:
|
||||
workflow_dispatch:
|
||||
|
||||
jobs:
|
||||
pr-title:
|
||||
if: github.event_name == 'pull_request'
|
||||
runs-on: ubuntu-latest
|
||||
permissions:
|
||||
pull-requests: read
|
||||
|
||||
steps:
|
||||
- name: Validate PR title
|
||||
uses: amannn/action-semantic-pull-request@v6.1.1
|
||||
env:
|
||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||
with:
|
||||
requireScope: false
|
||||
types: |
|
||||
feat
|
||||
fix
|
||||
docs
|
||||
refactor
|
||||
chore
|
||||
test
|
||||
ci
|
||||
build
|
||||
perf
|
||||
revert
|
||||
|
||||
test:
|
||||
runs-on: ubuntu-latest
|
||||
|
||||
@@ -21,5 +46,8 @@ jobs:
|
||||
- name: Install dependencies
|
||||
run: bun install
|
||||
|
||||
- name: Validate release metadata
|
||||
run: bun run release:validate
|
||||
|
||||
- name: Run tests
|
||||
run: bun test
|
||||
|
||||
37
.github/workflows/publish.yml
vendored
37
.github/workflows/publish.yml
vendored
@@ -1,37 +0,0 @@
|
||||
name: Publish to npm
|
||||
|
||||
on:
|
||||
release:
|
||||
types: [published]
|
||||
|
||||
jobs:
|
||||
publish:
|
||||
runs-on: ubuntu-latest
|
||||
permissions:
|
||||
contents: read
|
||||
id-token: write
|
||||
|
||||
steps:
|
||||
- uses: actions/checkout@v6
|
||||
|
||||
- name: Setup Bun
|
||||
uses: oven-sh/setup-bun@v2
|
||||
with:
|
||||
bun-version: latest
|
||||
|
||||
- name: Install dependencies
|
||||
run: bun install
|
||||
|
||||
- name: Run tests
|
||||
run: bun test
|
||||
|
||||
- name: Setup Node.js for npm publish
|
||||
uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version: "20"
|
||||
registry-url: "https://registry.npmjs.org"
|
||||
|
||||
- name: Publish to npm
|
||||
run: npm publish --provenance --access public
|
||||
env:
|
||||
NODE_AUTH_TOKEN: ${{ secrets.NPM_TOKEN }}
|
||||
84
.github/workflows/release-pr.yml
vendored
Normal file
84
.github/workflows/release-pr.yml
vendored
Normal file
@@ -0,0 +1,84 @@
|
||||
name: Release PR
|
||||
|
||||
on:
|
||||
push:
|
||||
branches: [main]
|
||||
workflow_dispatch:
|
||||
|
||||
permissions:
|
||||
contents: write
|
||||
pull-requests: write
|
||||
issues: write
|
||||
|
||||
concurrency:
|
||||
group: release-pr-${{ github.ref }}
|
||||
cancel-in-progress: false
|
||||
|
||||
jobs:
|
||||
release-pr:
|
||||
runs-on: ubuntu-latest
|
||||
outputs:
|
||||
cli_release_created: ${{ steps.release.outputs.release_created }}
|
||||
cli_tag_name: ${{ steps.release.outputs.tag_name }}
|
||||
|
||||
steps:
|
||||
- uses: actions/checkout@v6
|
||||
with:
|
||||
fetch-depth: 0
|
||||
|
||||
- name: Setup Bun
|
||||
uses: oven-sh/setup-bun@v2
|
||||
with:
|
||||
bun-version: latest
|
||||
|
||||
- name: Install dependencies
|
||||
run: bun install --frozen-lockfile
|
||||
|
||||
- name: Validate release metadata scripts
|
||||
run: bun run release:validate
|
||||
|
||||
- name: Maintain release PR
|
||||
id: release
|
||||
uses: googleapis/release-please-action@v4.4.0
|
||||
with:
|
||||
token: ${{ secrets.GITHUB_TOKEN }}
|
||||
config-file: .github/release-please-config.json
|
||||
manifest-file: .github/.release-please-manifest.json
|
||||
skip-labeling: true
|
||||
|
||||
publish-cli:
|
||||
needs: release-pr
|
||||
if: needs.release-pr.outputs.cli_release_created == 'true'
|
||||
runs-on: ubuntu-latest
|
||||
permissions:
|
||||
contents: read
|
||||
id-token: write
|
||||
|
||||
concurrency:
|
||||
group: publish-${{ needs.release-pr.outputs.cli_tag_name }}
|
||||
cancel-in-progress: false
|
||||
|
||||
steps:
|
||||
- uses: actions/checkout@v6
|
||||
with:
|
||||
fetch-depth: 0
|
||||
ref: ${{ needs.release-pr.outputs.cli_tag_name }}
|
||||
|
||||
- name: Setup Bun
|
||||
uses: oven-sh/setup-bun@v2
|
||||
with:
|
||||
bun-version: latest
|
||||
|
||||
- name: Install dependencies
|
||||
run: bun install --frozen-lockfile
|
||||
|
||||
- name: Run tests
|
||||
run: bun test
|
||||
|
||||
- name: Setup Node.js for release
|
||||
uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version: "24"
|
||||
|
||||
- name: Publish package
|
||||
run: npm publish --provenance --access public
|
||||
94
.github/workflows/release-preview.yml
vendored
Normal file
94
.github/workflows/release-preview.yml
vendored
Normal file
@@ -0,0 +1,94 @@
|
||||
name: Release Preview
|
||||
|
||||
on:
|
||||
workflow_dispatch:
|
||||
inputs:
|
||||
title:
|
||||
description: "Conventional title to evaluate (defaults to the latest commit title on this ref)"
|
||||
required: false
|
||||
type: string
|
||||
cli_bump:
|
||||
description: "CLI bump override"
|
||||
required: false
|
||||
type: choice
|
||||
options: [auto, patch, minor, major]
|
||||
default: auto
|
||||
compound_engineering_bump:
|
||||
description: "compound-engineering bump override"
|
||||
required: false
|
||||
type: choice
|
||||
options: [auto, patch, minor, major]
|
||||
default: auto
|
||||
coding_tutor_bump:
|
||||
description: "coding-tutor bump override"
|
||||
required: false
|
||||
type: choice
|
||||
options: [auto, patch, minor, major]
|
||||
default: auto
|
||||
marketplace_bump:
|
||||
description: "marketplace bump override"
|
||||
required: false
|
||||
type: choice
|
||||
options: [auto, patch, minor, major]
|
||||
default: auto
|
||||
|
||||
jobs:
|
||||
preview:
|
||||
runs-on: ubuntu-latest
|
||||
|
||||
steps:
|
||||
- uses: actions/checkout@v6
|
||||
with:
|
||||
fetch-depth: 0
|
||||
|
||||
- name: Setup Bun
|
||||
uses: oven-sh/setup-bun@v2
|
||||
with:
|
||||
bun-version: latest
|
||||
|
||||
- name: Install dependencies
|
||||
run: bun install --frozen-lockfile
|
||||
|
||||
- name: Determine title and changed files
|
||||
id: inputs
|
||||
shell: bash
|
||||
run: |
|
||||
TITLE="${{ github.event.inputs.title }}"
|
||||
if [ -z "$TITLE" ]; then
|
||||
TITLE="$(git log -1 --pretty=%s)"
|
||||
fi
|
||||
|
||||
FILES="$(git diff --name-only HEAD~1...HEAD | tr '\n' ' ')"
|
||||
|
||||
echo "title=$TITLE" >> "$GITHUB_OUTPUT"
|
||||
echo "files=$FILES" >> "$GITHUB_OUTPUT"
|
||||
|
||||
- name: Add preview note
|
||||
run: |
|
||||
echo "This preview currently evaluates the selected ref from its latest commit title and changed files." >> "$GITHUB_STEP_SUMMARY"
|
||||
echo "It is side-effect free, but it does not yet reconstruct the full accumulated open release PR state." >> "$GITHUB_STEP_SUMMARY"
|
||||
|
||||
- name: Validate release metadata
|
||||
run: bun run release:validate
|
||||
|
||||
- name: Preview release
|
||||
shell: bash
|
||||
run: |
|
||||
TITLE='${{ steps.inputs.outputs.title }}'
|
||||
FILES='${{ steps.inputs.outputs.files }}'
|
||||
|
||||
args=(--title "$TITLE" --json)
|
||||
for file in $FILES; do
|
||||
args+=(--file "$file")
|
||||
done
|
||||
|
||||
args+=(--override "cli=${{ github.event.inputs.cli_bump || 'auto' }}")
|
||||
args+=(--override "compound-engineering=${{ github.event.inputs.compound_engineering_bump || 'auto' }}")
|
||||
args+=(--override "coding-tutor=${{ github.event.inputs.coding_tutor_bump || 'auto' }}")
|
||||
args+=(--override "marketplace=${{ github.event.inputs.marketplace_bump || 'auto' }}")
|
||||
|
||||
bun run scripts/release/preview.ts "${args[@]}" | tee /tmp/release-preview.txt
|
||||
|
||||
- name: Publish preview summary
|
||||
shell: bash
|
||||
run: cat /tmp/release-preview.txt >> "$GITHUB_STEP_SUMMARY"
|
||||
1
.gitignore
vendored
1
.gitignore
vendored
@@ -3,3 +3,4 @@
|
||||
node_modules/
|
||||
.codex/
|
||||
todos/
|
||||
.worktrees
|
||||
|
||||
95
AGENTS.md
95
AGENTS.md
@@ -1,18 +1,85 @@
|
||||
# Agent Instructions
|
||||
|
||||
This repository contains a Bun/TypeScript CLI that converts Claude Code plugins into other agent platform formats.
|
||||
This repository primarily houses the `compound-engineering` coding-agent plugin and the Claude Code marketplace/catalog metadata used to distribute it.
|
||||
|
||||
It also contains:
|
||||
- the Bun/TypeScript CLI that converts Claude Code plugins into other agent platform formats
|
||||
- additional plugins under `plugins/`, such as `coding-tutor`
|
||||
- shared release and metadata infrastructure for the CLI, marketplace, and plugins
|
||||
|
||||
`AGENTS.md` is the canonical repo instruction file. Root `CLAUDE.md` exists only as a compatibility shim for tools and conversions that still look for it.
|
||||
|
||||
## Quick Start
|
||||
|
||||
```bash
|
||||
bun install
|
||||
bun test # full test suite
|
||||
bun run release:validate # check plugin/marketplace consistency
|
||||
```
|
||||
|
||||
## Working Agreement
|
||||
|
||||
- **Branching:** Create a feature branch for any non-trivial change. If already on the correct branch for the task, keep using it; do not create additional branches or worktrees unless explicitly requested.
|
||||
- **Safety:** Do not delete or overwrite user data. Avoid destructive commands.
|
||||
- **Testing:** Run `bun test` after changes that affect parsing, conversion, or output.
|
||||
- **Output Paths:** Keep OpenCode output at `opencode.json` and `.opencode/{agents,skills,plugins}`.
|
||||
- **Release versioning:** Releases are prepared by release automation, not normal feature PRs. The repo now has multiple release components (`cli`, `compound-engineering`, `coding-tutor`, `marketplace`). GitHub release PRs and GitHub Releases are the canonical release-notes surface for new releases; root `CHANGELOG.md` is only a pointer to that history. Use conventional titles such as `feat:` and `fix:` so release automation can classify change intent, but do not hand-bump release-owned versions or hand-author release notes in routine PRs.
|
||||
- **Output Paths:** Keep OpenCode output at `opencode.json` and `.opencode/{agents,skills,plugins}`. For OpenCode, command go to `~/.config/opencode/commands/<name>.md`; `opencode.json` is deep-merged (never overwritten wholesale).
|
||||
- **ASCII-first:** Use ASCII unless the file already contains Unicode.
|
||||
|
||||
## Adding a New Target Provider (e.g., Codex)
|
||||
## Directory Layout
|
||||
|
||||
Use this checklist when introducing a new target provider:
|
||||
```
|
||||
src/ CLI entry point, parsers, converters, target writers
|
||||
plugins/ Plugin workspaces (compound-engineering, coding-tutor)
|
||||
.claude-plugin/ Claude marketplace catalog metadata
|
||||
tests/ Converter, writer, and CLI tests + fixtures
|
||||
docs/ Requirements, plans, solutions, and target specs
|
||||
```
|
||||
|
||||
## Repo Surfaces
|
||||
|
||||
Changes in this repo may affect one or more of these surfaces:
|
||||
|
||||
- `compound-engineering` under `plugins/compound-engineering/`
|
||||
- the Claude marketplace catalog under `.claude-plugin/`
|
||||
- the converter/install CLI in `src/` and `package.json`
|
||||
- secondary plugins such as `plugins/coding-tutor/`
|
||||
|
||||
Do not assume a repo change is "just CLI" or "just plugin" without checking which surface owns the affected files.
|
||||
|
||||
## Plugin Maintenance
|
||||
|
||||
When changing `plugins/compound-engineering/` content:
|
||||
|
||||
- Update substantive docs like `plugins/compound-engineering/README.md` when the plugin behavior, inventory, or usage changes.
|
||||
- Do not hand-bump release-owned versions in plugin or marketplace manifests.
|
||||
- Do not hand-add release entries to `CHANGELOG.md` or treat it as the canonical source for new releases.
|
||||
- Run `bun run release:validate` if agents, commands, skills, MCP servers, or release-owned descriptions/counts may have changed.
|
||||
|
||||
Useful validation commands:
|
||||
|
||||
```bash
|
||||
bun run release:validate
|
||||
cat .claude-plugin/marketplace.json | jq .
|
||||
cat plugins/compound-engineering/.claude-plugin/plugin.json | jq .
|
||||
```
|
||||
|
||||
## Coding Conventions
|
||||
|
||||
- Prefer explicit mappings over implicit magic when converting between platforms.
|
||||
- Keep target-specific behavior in dedicated converters/writers instead of scattering conditionals across unrelated files.
|
||||
- Preserve stable output paths and merge semantics for installed targets; do not casually change generated file locations.
|
||||
- When adding or changing a target, update fixtures/tests alongside implementation rather than treating docs or examples as sufficient proof.
|
||||
|
||||
## Commit Conventions
|
||||
|
||||
- Use conventional titles such as `feat: ...`, `fix: ...`, `docs: ...`, and `refactor: ...`.
|
||||
- Component scope is optional. Example: `feat(coding-tutor): add quiz reset`.
|
||||
- Breaking changes must be explicit with `!` or a breaking-change footer so release automation can classify them correctly.
|
||||
|
||||
## Adding a New Target Provider
|
||||
|
||||
Only add a provider when the target format is stable, documented, and has a clear mapping for tools/permissions/hooks. Use this checklist:
|
||||
|
||||
1. **Define the target entry**
|
||||
- Add a new handler in `src/targets/index.ts` with `implemented: false` until complete.
|
||||
@@ -36,13 +103,19 @@ Use this checklist when introducing a new target provider:
|
||||
5. **Docs**
|
||||
- Update README with the new `--to` option and output locations.
|
||||
|
||||
## When to Add a Provider
|
||||
## Agent References in Skills
|
||||
|
||||
Add a new provider when at least one of these is true:
|
||||
When referencing agents from within skill SKILL.md files (e.g., via the `Agent` or `Task` tool), always use the **fully-qualified namespace**: `compound-engineering:<category>:<agent-name>`. Never use the short agent name alone.
|
||||
|
||||
- A real user/workflow needs it now.
|
||||
- The target format is stable and documented.
|
||||
- There’s a clear mapping for tools/permissions/hooks.
|
||||
- You can write fixtures + tests that validate the mapping.
|
||||
Example:
|
||||
- `compound-engineering:research:learnings-researcher` (correct)
|
||||
- `learnings-researcher` (wrong - will fail to resolve at runtime)
|
||||
|
||||
Avoid adding a provider if the target spec is unstable or undocumented.
|
||||
This prevents resolution failures when the plugin is installed alongside other plugins that may define agents with the same short name.
|
||||
|
||||
## Repository Docs Convention
|
||||
|
||||
- **Requirements** live in `docs/brainstorms/` — requirements exploration and ideation.
|
||||
- **Plans** live in `docs/plans/` — implementation plans and progress tracking.
|
||||
- **Solutions** live in `docs/solutions/` — documented decisions and patterns.
|
||||
- **Specs** live in `docs/specs/` — target platform format specifications.
|
||||
|
||||
36
CHANGELOG.md
36
CHANGELOG.md
@@ -1,34 +1,14 @@
|
||||
# Changelog
|
||||
|
||||
All notable changes to the `@every-env/compound-plugin` CLI tool will be documented in this file.
|
||||
Release notes now live in GitHub Releases for this repository:
|
||||
|
||||
The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.0.0/),
|
||||
and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.html).
|
||||
https://github.com/EveryInc/compound-engineering-plugin/releases
|
||||
|
||||
## [0.6.0] - 2026-02-12
|
||||
Multi-component releases are published under component-specific tags such as:
|
||||
|
||||
### Added
|
||||
- `cli-vX.Y.Z`
|
||||
- `compound-engineering-vX.Y.Z`
|
||||
- `coding-tutor-vX.Y.Z`
|
||||
- `marketplace-vX.Y.Z`
|
||||
|
||||
- **Droid sync target** — `sync --target droid` symlinks personal skills to `~/.factory/skills/`
|
||||
- **Cursor sync target** — `sync --target cursor` symlinks skills to `.cursor/skills/` and merges MCP servers into `.cursor/mcp.json`
|
||||
- **Pi target** — First-class `--to pi` converter with MCPorter config and subagent compatibility ([#181](https://github.com/EveryInc/compound-engineering-plugin/pull/181)) — thanks [@gvkhosla](https://github.com/gvkhosla)!
|
||||
|
||||
### Fixed
|
||||
|
||||
- **Bare Claude model alias resolution** — Fixed OpenCode converter not resolving bare model aliases like `claude-sonnet-4-5-20250514` ([#182](https://github.com/EveryInc/compound-engineering-plugin/pull/182)) — thanks [@waltbeaman](https://github.com/waltbeaman)!
|
||||
|
||||
### Changed
|
||||
|
||||
- Extracted shared `expandHome` / `resolveTargetHome` helpers to `src/utils/resolve-home.ts`, removing duplication across `convert.ts`, `install.ts`, and `sync.ts`
|
||||
|
||||
---
|
||||
|
||||
## [0.5.2] - 2026-02-09
|
||||
|
||||
### Fixed
|
||||
|
||||
- Fix cursor install defaulting to cwd instead of opencode config dir
|
||||
|
||||
## [0.5.1] - 2026-02-08
|
||||
|
||||
- Initial npm publish
|
||||
Do not add new release entries here. New release notes are managed by release automation in GitHub.
|
||||
|
||||
381
CLAUDE.md
381
CLAUDE.md
@@ -1,380 +1 @@
|
||||
# Every Marketplace - Claude Code Plugin Marketplace
|
||||
|
||||
This repository is a Claude Code plugin marketplace that distributes the `compound-engineering` plugin to developers building with AI-powered tools.
|
||||
|
||||
## Repository Structure
|
||||
|
||||
```
|
||||
every-marketplace/
|
||||
├── .claude-plugin/
|
||||
│ └── marketplace.json # Marketplace catalog (lists available plugins)
|
||||
├── docs/ # Documentation site (GitHub Pages)
|
||||
│ ├── index.html # Landing page
|
||||
│ ├── css/ # Stylesheets
|
||||
│ ├── js/ # JavaScript
|
||||
│ └── pages/ # Reference pages
|
||||
└── plugins/
|
||||
└── compound-engineering/ # The actual plugin
|
||||
├── .claude-plugin/
|
||||
│ └── plugin.json # Plugin metadata
|
||||
├── agents/ # 24 specialized AI agents
|
||||
├── commands/ # 13 slash commands
|
||||
├── skills/ # 11 skills
|
||||
├── mcp-servers/ # 2 MCP servers (playwright, context7)
|
||||
├── README.md # Plugin documentation
|
||||
└── CHANGELOG.md # Version history
|
||||
```
|
||||
|
||||
## Philosophy: Compounding Engineering
|
||||
|
||||
**Each unit of engineering work should make subsequent units of work easier—not harder.**
|
||||
|
||||
When working on this repository, follow the compounding engineering process:
|
||||
|
||||
1. **Plan** → Understand the change needed and its impact
|
||||
2. **Delegate** → Use AI tools to help with implementation
|
||||
3. **Assess** → Verify changes work as expected
|
||||
4. **Codify** → Update this CLAUDE.md with learnings
|
||||
|
||||
## Working with This Repository
|
||||
|
||||
### Adding a New Plugin
|
||||
|
||||
1. Create plugin directory: `plugins/new-plugin-name/`
|
||||
2. Add plugin structure:
|
||||
```
|
||||
plugins/new-plugin-name/
|
||||
├── .claude-plugin/plugin.json
|
||||
├── agents/
|
||||
├── commands/
|
||||
└── README.md
|
||||
```
|
||||
3. Update `.claude-plugin/marketplace.json` to include the new plugin
|
||||
4. Test locally before committing
|
||||
|
||||
### Updating the Compounding Engineering Plugin
|
||||
|
||||
When agents, commands, or skills are added/removed, follow this checklist:
|
||||
|
||||
#### 1. Count all components accurately
|
||||
|
||||
```bash
|
||||
# Count agents
|
||||
ls plugins/compound-engineering/agents/*.md | wc -l
|
||||
|
||||
# Count commands
|
||||
ls plugins/compound-engineering/commands/*.md | wc -l
|
||||
|
||||
# Count skills
|
||||
ls -d plugins/compound-engineering/skills/*/ 2>/dev/null | wc -l
|
||||
```
|
||||
|
||||
#### 2. Update ALL description strings with correct counts
|
||||
|
||||
The description appears in multiple places and must match everywhere:
|
||||
|
||||
- [ ] `plugins/compound-engineering/.claude-plugin/plugin.json` → `description` field
|
||||
- [ ] `.claude-plugin/marketplace.json` → plugin `description` field
|
||||
- [ ] `plugins/compound-engineering/README.md` → intro paragraph
|
||||
|
||||
Format: `"Includes X specialized agents, Y commands, and Z skill(s)."`
|
||||
|
||||
#### 3. Update version numbers
|
||||
|
||||
When adding new functionality, bump the version in:
|
||||
|
||||
- [ ] `plugins/compound-engineering/.claude-plugin/plugin.json` → `version`
|
||||
- [ ] `.claude-plugin/marketplace.json` → plugin `version`
|
||||
|
||||
#### 4. Update documentation
|
||||
|
||||
- [ ] `plugins/compound-engineering/README.md` → list all components
|
||||
- [ ] `plugins/compound-engineering/CHANGELOG.md` → document changes
|
||||
- [ ] `CLAUDE.md` → update structure diagram if needed
|
||||
|
||||
#### 5. Rebuild documentation site
|
||||
|
||||
Run the release-docs command to update all documentation pages:
|
||||
|
||||
```bash
|
||||
claude /release-docs
|
||||
```
|
||||
|
||||
This will:
|
||||
- Update stats on the landing page
|
||||
- Regenerate reference pages (agents, commands, skills, MCP servers)
|
||||
- Update the changelog page
|
||||
- Validate all counts match actual files
|
||||
|
||||
#### 6. Validate JSON files
|
||||
|
||||
```bash
|
||||
cat .claude-plugin/marketplace.json | jq .
|
||||
cat plugins/compound-engineering/.claude-plugin/plugin.json | jq .
|
||||
```
|
||||
|
||||
#### 6. Verify before committing
|
||||
|
||||
```bash
|
||||
# Ensure counts in descriptions match actual files
|
||||
grep -o "Includes [0-9]* specialized agents" plugins/compound-engineering/.claude-plugin/plugin.json
|
||||
ls plugins/compound-engineering/agents/*.md | wc -l
|
||||
```
|
||||
|
||||
### Marketplace.json Structure
|
||||
|
||||
The marketplace.json follows the official Claude Code spec:
|
||||
|
||||
```json
|
||||
{
|
||||
"name": "marketplace-identifier",
|
||||
"owner": {
|
||||
"name": "Owner Name",
|
||||
"url": "https://github.com/owner"
|
||||
},
|
||||
"metadata": {
|
||||
"description": "Marketplace description",
|
||||
"version": "1.0.0"
|
||||
},
|
||||
"plugins": [
|
||||
{
|
||||
"name": "plugin-name",
|
||||
"description": "Plugin description",
|
||||
"version": "1.0.0",
|
||||
"author": { ... },
|
||||
"homepage": "https://...",
|
||||
"tags": ["tag1", "tag2"],
|
||||
"source": "./plugins/plugin-name"
|
||||
}
|
||||
]
|
||||
}
|
||||
```
|
||||
|
||||
**Only include fields that are in the official spec.** Do not add custom fields like:
|
||||
|
||||
- `downloads`, `stars`, `rating` (display-only)
|
||||
- `categories`, `featured_plugins`, `trending` (not in spec)
|
||||
- `type`, `verified`, `featured` (not in spec)
|
||||
|
||||
### Plugin.json Structure
|
||||
|
||||
Each plugin has its own plugin.json with detailed metadata:
|
||||
|
||||
```json
|
||||
{
|
||||
"name": "plugin-name",
|
||||
"version": "1.0.0",
|
||||
"description": "Plugin description",
|
||||
"author": { ... },
|
||||
"keywords": ["keyword1", "keyword2"],
|
||||
"components": {
|
||||
"agents": 15,
|
||||
"commands": 6,
|
||||
"hooks": 2
|
||||
},
|
||||
"agents": {
|
||||
"category": [
|
||||
{
|
||||
"name": "agent-name",
|
||||
"description": "Agent description",
|
||||
"use_cases": ["use-case-1", "use-case-2"]
|
||||
}
|
||||
]
|
||||
},
|
||||
"commands": {
|
||||
"category": ["command1", "command2"]
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
## Documentation Site
|
||||
|
||||
The documentation site is at `/docs` in the repository root (for GitHub Pages). This site is built with plain HTML/CSS/JS (based on Evil Martians' LaunchKit template) and requires no build step to view.
|
||||
|
||||
### Documentation Structure
|
||||
|
||||
```
|
||||
docs/
|
||||
├── index.html # Landing page with stats and philosophy
|
||||
├── css/
|
||||
│ ├── style.css # Main styles (LaunchKit-based)
|
||||
│ └── docs.css # Documentation-specific styles
|
||||
├── js/
|
||||
│ └── main.js # Interactivity (theme toggle, mobile nav)
|
||||
└── pages/
|
||||
├── getting-started.html # Installation and quick start
|
||||
├── agents.html # All 24 agents reference
|
||||
├── commands.html # All 13 commands reference
|
||||
├── skills.html # All 11 skills reference
|
||||
├── mcp-servers.html # MCP servers reference
|
||||
└── changelog.html # Version history
|
||||
```
|
||||
|
||||
### Keeping Docs Up-to-Date
|
||||
|
||||
**IMPORTANT:** After ANY change to agents, commands, skills, or MCP servers, run:
|
||||
|
||||
```bash
|
||||
claude /release-docs
|
||||
```
|
||||
|
||||
This command:
|
||||
1. Counts all current components
|
||||
2. Reads all agent/command/skill/MCP files
|
||||
3. Regenerates all reference pages
|
||||
4. Updates stats on the landing page
|
||||
5. Updates the changelog from CHANGELOG.md
|
||||
6. Validates counts match across all files
|
||||
|
||||
### Manual Updates
|
||||
|
||||
If you need to update docs manually:
|
||||
|
||||
1. **Landing page stats** - Update the numbers in `docs/index.html`:
|
||||
```html
|
||||
<span class="stat-number">24</span> <!-- agents -->
|
||||
<span class="stat-number">13</span> <!-- commands -->
|
||||
```
|
||||
|
||||
2. **Reference pages** - Each page in `docs/pages/` documents all components in that category
|
||||
|
||||
3. **Changelog** - `docs/pages/changelog.html` mirrors `CHANGELOG.md` in HTML format
|
||||
|
||||
### Viewing Docs Locally
|
||||
|
||||
Since the docs are static HTML, you can view them directly:
|
||||
|
||||
```bash
|
||||
# Open in browser
|
||||
open docs/index.html
|
||||
|
||||
# Or start a local server
|
||||
cd docs
|
||||
python -m http.server 8000
|
||||
# Then visit http://localhost:8000
|
||||
```
|
||||
|
||||
## Testing Changes
|
||||
|
||||
### Test Locally
|
||||
|
||||
1. Install the marketplace locally:
|
||||
|
||||
```bash
|
||||
claude /plugin marketplace add /Users/yourusername/every-marketplace
|
||||
```
|
||||
|
||||
2. Install the plugin:
|
||||
|
||||
```bash
|
||||
claude /plugin install compound-engineering
|
||||
```
|
||||
|
||||
3. Test agents and commands:
|
||||
```bash
|
||||
claude /review
|
||||
claude agent kieran-rails-reviewer "test message"
|
||||
```
|
||||
|
||||
### Validate JSON
|
||||
|
||||
Before committing, ensure JSON files are valid:
|
||||
|
||||
```bash
|
||||
cat .claude-plugin/marketplace.json | jq .
|
||||
cat plugins/compound-engineering/.claude-plugin/plugin.json | jq .
|
||||
```
|
||||
|
||||
## Common Tasks
|
||||
|
||||
### Adding a New Agent
|
||||
|
||||
1. Create `plugins/compound-engineering/agents/new-agent.md`
|
||||
2. Update plugin.json agent count and agent list
|
||||
3. Update README.md agent list
|
||||
4. Test with `claude agent new-agent "test"`
|
||||
|
||||
### Adding a New Command
|
||||
|
||||
1. Create `plugins/compound-engineering/commands/new-command.md`
|
||||
2. Update plugin.json command count and command list
|
||||
3. Update README.md command list
|
||||
4. Test with `claude /new-command`
|
||||
|
||||
### Adding a New Skill
|
||||
|
||||
1. Create skill directory: `plugins/compound-engineering/skills/skill-name/`
|
||||
2. Add skill structure:
|
||||
```
|
||||
skills/skill-name/
|
||||
├── SKILL.md # Skill definition with frontmatter (name, description)
|
||||
└── scripts/ # Supporting scripts (optional)
|
||||
```
|
||||
3. Update plugin.json description with new skill count
|
||||
4. Update marketplace.json description with new skill count
|
||||
5. Update README.md with skill documentation
|
||||
6. Update CHANGELOG.md with the addition
|
||||
7. Test with `claude skill skill-name`
|
||||
|
||||
**Skill file format (SKILL.md):**
|
||||
```markdown
|
||||
---
|
||||
name: skill-name
|
||||
description: Brief description of what the skill does
|
||||
---
|
||||
|
||||
# Skill Title
|
||||
|
||||
Detailed documentation...
|
||||
```
|
||||
|
||||
### Updating Tags/Keywords
|
||||
|
||||
Tags should reflect the compounding engineering philosophy:
|
||||
|
||||
- Use: `ai-powered`, `compound-engineering`, `workflow-automation`, `knowledge-management`
|
||||
- Avoid: Framework-specific tags unless the plugin is framework-specific
|
||||
|
||||
## Commit Conventions
|
||||
|
||||
Follow these patterns for commit messages:
|
||||
|
||||
- `Add [agent/command name]` - Adding new functionality
|
||||
- `Remove [agent/command name]` - Removing functionality
|
||||
- `Update [file] to [what changed]` - Updating existing files
|
||||
- `Fix [issue]` - Bug fixes
|
||||
- `Simplify [component] to [improvement]` - Refactoring
|
||||
|
||||
Include the Claude Code footer:
|
||||
|
||||
```
|
||||
🤖 Generated with [Claude Code](https://claude.com/claude-code)
|
||||
|
||||
Co-Authored-By: Claude <noreply@anthropic.com>
|
||||
```
|
||||
|
||||
## Resources to search for when needing more information
|
||||
|
||||
- [Claude Code Plugin Documentation](https://docs.claude.com/en/docs/claude-code/plugins)
|
||||
- [Plugin Marketplace Documentation](https://docs.claude.com/en/docs/claude-code/plugin-marketplaces)
|
||||
- [Plugin Reference](https://docs.claude.com/en/docs/claude-code/plugins-reference)
|
||||
|
||||
## Key Learnings
|
||||
|
||||
_This section captures important learnings as we work on this repository._
|
||||
|
||||
### 2024-11-22: Added gemini-imagegen skill and fixed component counts
|
||||
|
||||
Added the first skill to the plugin and discovered the component counts were wrong (said 15 agents, actually had 17). Created a comprehensive checklist for updating the plugin to prevent this in the future.
|
||||
|
||||
**Learning:** Always count actual files before updating descriptions. The counts appear in multiple places (plugin.json, marketplace.json, README.md) and must all match. Use the verification commands in the checklist above.
|
||||
|
||||
### 2024-10-09: Simplified marketplace.json to match official spec
|
||||
|
||||
The initial marketplace.json included many custom fields (downloads, stars, rating, categories, trending) that aren't part of the Claude Code specification. We simplified to only include:
|
||||
|
||||
- Required: `name`, `owner`, `plugins`
|
||||
- Optional: `metadata` (with description and version)
|
||||
- Plugin entries: `name`, `description`, `version`, `author`, `homepage`, `tags`, `source`
|
||||
|
||||
**Learning:** Stick to the official spec. Custom fields may confuse users or break compatibility with future versions.
|
||||
@AGENTS.md
|
||||
|
||||
38
PRIVACY.md
Normal file
38
PRIVACY.md
Normal file
@@ -0,0 +1,38 @@
|
||||
# Privacy & Data Handling
|
||||
|
||||
This repository contains:
|
||||
- a plugin package (`plugins/compound-engineering`) made of markdown/config content
|
||||
- a CLI (`@every-env/compound-plugin`) that converts and installs plugin content for different AI coding tools
|
||||
|
||||
## Summary
|
||||
|
||||
- The plugin package does not include telemetry or analytics code.
|
||||
- The plugin package does not run a background service that uploads repository/workspace contents automatically.
|
||||
- Data leaves your machine only when your host/tooling or an explicitly invoked integration performs a network request.
|
||||
|
||||
## What May Send Data
|
||||
|
||||
1. AI host/model providers
|
||||
|
||||
If you run the plugin in tools like Claude Code, Cursor, Gemini CLI, Copilot, Kiro, Windsurf, etc., those tools may send prompts/context/code to their configured model providers. This behavior is controlled by those tools and providers, not by this plugin repository.
|
||||
|
||||
2. Optional integrations and tools
|
||||
|
||||
The plugin includes optional capabilities that can call external services when explicitly used, for example:
|
||||
- Context7 MCP (`https://mcp.context7.com/mcp`) for documentation lookup
|
||||
- Proof (`https://www.proofeditor.ai`) when using share/edit flows
|
||||
- Other opt-in skills (for example image generation or cloud upload workflows) that call their own external APIs/services
|
||||
|
||||
If you do not invoke these integrations, they do not transmit your project data.
|
||||
|
||||
3. Package/installer infrastructure
|
||||
|
||||
Installing dependencies or packages (for example `npm`, `bunx`) communicates with package registries/CDNs according to your package manager configuration.
|
||||
|
||||
## Data Ownership and Retention
|
||||
|
||||
This repository does not operate a backend service for collecting or storing your project/workspace data. Data retention and processing for model prompts or optional integrations are governed by the external services you use.
|
||||
|
||||
## Security Reporting
|
||||
|
||||
If you identify a security issue in this repository, follow the disclosure process in [SECURITY.md](SECURITY.md).
|
||||
156
README.md
156
README.md
@@ -8,13 +8,19 @@ A Claude Code plugin marketplace featuring the **Compound Engineering Plugin**
|
||||
## Claude Code Install
|
||||
|
||||
```bash
|
||||
/plugin marketplace add https://github.com/EveryInc/compound-engineering-plugin
|
||||
/plugin marketplace add EveryInc/compound-engineering-plugin
|
||||
/plugin install compound-engineering
|
||||
```
|
||||
|
||||
## OpenCode, Codex, Droid, Cursor, Pi & Gemini (experimental) Install
|
||||
## Cursor Install
|
||||
|
||||
This repo includes a Bun/TypeScript CLI that converts Claude Code plugins to OpenCode, Codex, Factory Droid, Cursor, Pi, and Gemini CLI.
|
||||
```text
|
||||
/add-plugin compound-engineering
|
||||
```
|
||||
|
||||
## OpenCode, Codex, Droid, Pi, Gemini, Copilot, Kiro, Windsurf, OpenClaw & Qwen (experimental) Install
|
||||
|
||||
This repo includes a Bun/TypeScript CLI that converts Claude Code plugins to OpenCode, Codex, Factory Droid, Pi, Gemini CLI, GitHub Copilot, Kiro CLI, Windsurf, OpenClaw, and Qwen Code.
|
||||
|
||||
```bash
|
||||
# convert the compound-engineering plugin into OpenCode format
|
||||
@@ -26,36 +32,93 @@ bunx @every-env/compound-plugin install compound-engineering --to codex
|
||||
# convert to Factory Droid format
|
||||
bunx @every-env/compound-plugin install compound-engineering --to droid
|
||||
|
||||
# convert to Cursor format
|
||||
bunx @every-env/compound-plugin install compound-engineering --to cursor
|
||||
|
||||
# convert to Pi format
|
||||
bunx @every-env/compound-plugin install compound-engineering --to pi
|
||||
|
||||
# convert to Gemini CLI format
|
||||
bunx @every-env/compound-plugin install compound-engineering --to gemini
|
||||
|
||||
# convert to GitHub Copilot format
|
||||
bunx @every-env/compound-plugin install compound-engineering --to copilot
|
||||
|
||||
# convert to Kiro CLI format
|
||||
bunx @every-env/compound-plugin install compound-engineering --to kiro
|
||||
|
||||
# convert to OpenClaw format
|
||||
bunx @every-env/compound-plugin install compound-engineering --to openclaw
|
||||
|
||||
# convert to Windsurf format (global scope by default)
|
||||
bunx @every-env/compound-plugin install compound-engineering --to windsurf
|
||||
|
||||
# convert to Windsurf workspace scope
|
||||
bunx @every-env/compound-plugin install compound-engineering --to windsurf --scope workspace
|
||||
|
||||
# convert to Qwen Code format
|
||||
bunx @every-env/compound-plugin install compound-engineering --to qwen
|
||||
|
||||
# auto-detect installed tools and install to all
|
||||
bunx @every-env/compound-plugin install compound-engineering --to all
|
||||
```
|
||||
|
||||
Local dev:
|
||||
### Local Development
|
||||
|
||||
When developing and testing local changes to the plugin:
|
||||
|
||||
**Claude Code** — add a shell alias so your local copy loads alongside your normal plugins:
|
||||
|
||||
```bash
|
||||
# add to ~/.zshrc or ~/.bashrc
|
||||
alias claude-dev-ce='claude --plugin-dir ~/code/compound-engineering-plugin/plugins/compound-engineering'
|
||||
```
|
||||
|
||||
One-liner to append it:
|
||||
|
||||
```bash
|
||||
echo "alias claude-dev-ce='claude --plugin-dir ~/code/compound-engineering-plugin/plugins/compound-engineering'" >> ~/.zshrc
|
||||
```
|
||||
|
||||
Then run `claude-dev-ce` instead of `claude` to test your changes. Your production install stays untouched.
|
||||
|
||||
**Codex** — point the install command at your local path:
|
||||
|
||||
```bash
|
||||
bun run src/index.ts install ./plugins/compound-engineering --to codex
|
||||
```
|
||||
|
||||
**Other targets** — same pattern, swap the target:
|
||||
|
||||
```bash
|
||||
bun run src/index.ts install ./plugins/compound-engineering --to opencode
|
||||
```
|
||||
|
||||
OpenCode output is written to `~/.config/opencode` by default, with `opencode.json` at the root and `agents/`, `skills/`, and `plugins/` alongside it.
|
||||
Codex output is written to `~/.codex/prompts` and `~/.codex/skills`, with each Claude command converted into both a prompt and a skill (the prompt instructs Codex to load the corresponding skill). Generated Codex skill descriptions are truncated to 1024 characters (Codex limit).
|
||||
Droid output is written to `~/.factory/` with commands, droids (agents), and skills. Claude tool names are mapped to Factory equivalents (`Bash` → `Execute`, `Write` → `Create`, etc.) and namespace prefixes are stripped from commands.
|
||||
Cursor output is written to `.cursor/` with rules (`.mdc`), commands, skills, and `mcp.json`. Agents become "Agent Requested" rules (`alwaysApply: false`) so Cursor's AI activates them on demand. Works with both the Cursor IDE and Cursor CLI (`cursor-agent`) — they share the same `.cursor/` config directory.
|
||||
Pi output is written to `~/.pi/agent/` by default with prompts, skills, extensions, and `compound-engineering/mcporter.json` for MCPorter interoperability.
|
||||
Gemini output is written to `.gemini/` with skills (from agents), commands (`.toml`), and `settings.json` (MCP servers). Namespaced commands create directory structure (`workflows:plan` → `commands/workflows/plan.toml`). Skills use the identical SKILL.md standard and pass through unchanged.
|
||||
<details>
|
||||
<summary>Output format details per target</summary>
|
||||
|
||||
| Target | Output path | Notes |
|
||||
|--------|------------|-------|
|
||||
| `opencode` | `~/.config/opencode/` | Commands as `.md` files; `opencode.json` MCP config deep-merged; backups made before overwriting |
|
||||
| `codex` | `~/.codex/prompts` + `~/.codex/skills` | Claude commands become prompt + skill pairs; canonical `ce:*` workflow skills also get prompt wrappers; deprecated `workflows:*` aliases are omitted |
|
||||
| `droid` | `~/.factory/` | Tool names mapped (`Bash`→`Execute`, `Write`→`Create`); namespace prefixes stripped |
|
||||
| `pi` | `~/.pi/agent/` | Prompts, skills, extensions, and `mcporter.json` for MCPorter interoperability |
|
||||
| `gemini` | `.gemini/` | Skills from agents; commands as `.toml`; namespaced commands become directories (`workflows:plan` → `commands/workflows/plan.toml`) |
|
||||
| `copilot` | `.github/` | Agents as `.agent.md` with Copilot frontmatter; MCP env vars prefixed with `COPILOT_MCP_` |
|
||||
| `kiro` | `.kiro/` | Agents as JSON configs + prompt `.md` files; only stdio MCP servers supported |
|
||||
| `openclaw` | `~/.openclaw/extensions/<plugin>/` | Entry-point TypeScript skill file; `openclaw-extension.json` for MCP servers |
|
||||
| `windsurf` | `~/.codeium/windsurf/` (global) or `.windsurf/` (workspace) | Agents become skills; commands become flat workflows; `mcp_config.json` merged |
|
||||
| `qwen` | `~/.qwen/extensions/<plugin>/` | Agents as `.yaml`; env vars with placeholders extracted as settings; colon separator for nested commands |
|
||||
|
||||
All provider targets are experimental and may change as the formats evolve.
|
||||
|
||||
</details>
|
||||
|
||||
## Sync Personal Config
|
||||
|
||||
Sync your personal Claude Code config (`~/.claude/`) to other AI coding tools:
|
||||
Sync your personal Claude Code config (`~/.claude/`) to other AI coding tools. Omit `--target` to sync to all detected supported tools automatically:
|
||||
|
||||
```bash
|
||||
# Sync to all detected tools (default)
|
||||
bunx @every-env/compound-plugin sync
|
||||
|
||||
# Sync skills and MCP servers to OpenCode
|
||||
bunx @every-env/compound-plugin sync --target opencode
|
||||
|
||||
@@ -65,33 +128,80 @@ bunx @every-env/compound-plugin sync --target codex
|
||||
# Sync to Pi
|
||||
bunx @every-env/compound-plugin sync --target pi
|
||||
|
||||
# Sync to Droid (skills only)
|
||||
# Sync to Droid
|
||||
bunx @every-env/compound-plugin sync --target droid
|
||||
|
||||
# Sync to Cursor (skills + MCP servers)
|
||||
bunx @every-env/compound-plugin sync --target cursor
|
||||
# Sync to GitHub Copilot (skills + MCP servers)
|
||||
bunx @every-env/compound-plugin sync --target copilot
|
||||
|
||||
# Sync to Gemini (skills + MCP servers)
|
||||
bunx @every-env/compound-plugin sync --target gemini
|
||||
|
||||
# Sync to Windsurf
|
||||
bunx @every-env/compound-plugin sync --target windsurf
|
||||
|
||||
# Sync to Kiro
|
||||
bunx @every-env/compound-plugin sync --target kiro
|
||||
|
||||
# Sync to Qwen
|
||||
bunx @every-env/compound-plugin sync --target qwen
|
||||
|
||||
# Sync to OpenClaw (skills only; MCP is validation-gated)
|
||||
bunx @every-env/compound-plugin sync --target openclaw
|
||||
|
||||
# Sync to all detected tools
|
||||
bunx @every-env/compound-plugin sync --target all
|
||||
```
|
||||
|
||||
This syncs:
|
||||
- Personal skills from `~/.claude/skills/` (as symlinks)
|
||||
- Personal slash commands from `~/.claude/commands/` (as provider-native prompts, workflows, or converted skills where supported)
|
||||
- MCP servers from `~/.claude/settings.json`
|
||||
|
||||
Skills are symlinked (not copied) so changes in Claude Code are reflected immediately.
|
||||
|
||||
Supported sync targets:
|
||||
- `opencode`
|
||||
- `codex`
|
||||
- `pi`
|
||||
- `droid`
|
||||
- `copilot`
|
||||
- `gemini`
|
||||
- `windsurf`
|
||||
- `kiro`
|
||||
- `qwen`
|
||||
- `openclaw`
|
||||
|
||||
Notes:
|
||||
- Codex sync preserves non-managed `config.toml` content and now includes remote MCP servers.
|
||||
- Command sync reuses each provider's existing Claude command conversion, so some targets receive prompts or workflows while others receive converted skills.
|
||||
- Copilot sync writes personal skills to `~/.copilot/skills/` and MCP config to `~/.copilot/mcp-config.json`.
|
||||
- Gemini sync writes MCP config to `~/.gemini/` and avoids mirroring skills that Gemini already discovers from `~/.agents/skills`, which prevents duplicate-skill warnings.
|
||||
- Droid, Windsurf, Kiro, and Qwen sync merge MCP servers into the provider's documented user config.
|
||||
- OpenClaw currently syncs skills only. Personal command sync is skipped because this repo does not yet have a documented user-level OpenClaw command surface, and MCP sync is skipped because the current official OpenClaw docs do not clearly document an MCP server config contract.
|
||||
|
||||
## Workflow
|
||||
|
||||
```
|
||||
Plan → Work → Review → Compound → Repeat
|
||||
Brainstorm → Plan → Work → Review → Compound → Repeat
|
||||
↑
|
||||
Ideate (optional — when you need ideas)
|
||||
```
|
||||
|
||||
| Command | Purpose |
|
||||
|---------|---------|
|
||||
| `/workflows:plan` | Turn feature ideas into detailed implementation plans |
|
||||
| `/workflows:work` | Execute plans with worktrees and task tracking |
|
||||
| `/workflows:review` | Multi-agent code review before merging |
|
||||
| `/workflows:compound` | Document learnings to make future work easier |
|
||||
| `/ce:ideate` | Discover high-impact project improvements through divergent ideation and adversarial filtering |
|
||||
| `/ce:brainstorm` | Explore requirements and approaches before planning |
|
||||
| `/ce:plan` | Turn feature ideas into detailed implementation plans |
|
||||
| `/ce:work` | Execute plans with worktrees and task tracking |
|
||||
| `/ce:review` | Multi-agent code review before merging |
|
||||
| `/ce:compound` | Document learnings to make future work easier |
|
||||
|
||||
Each cycle compounds: plans inform future plans, reviews catch more issues, patterns get documented.
|
||||
The `/ce:ideate` skill proactively surfaces strong improvement ideas, and `/ce:brainstorm` then clarifies the selected one before committing to a plan.
|
||||
|
||||
Each cycle compounds: brainstorms sharpen plans, plans inform future plans, reviews catch more issues, patterns get documented.
|
||||
|
||||
> **Beta:** Experimental versions of `/ce:plan` and `/deepen-plan` are available as `/ce:plan-beta` and `/deepen-plan-beta`. See the [plugin README](plugins/compound-engineering/README.md#beta-skills) for details.
|
||||
|
||||
## Philosophy
|
||||
|
||||
|
||||
29
SECURITY.md
Normal file
29
SECURITY.md
Normal file
@@ -0,0 +1,29 @@
|
||||
# Security Policy
|
||||
|
||||
## Supported Versions
|
||||
|
||||
Security fixes are applied to the latest version on `main`.
|
||||
|
||||
## Reporting a Vulnerability
|
||||
|
||||
Please do not open a public issue for undisclosed vulnerabilities.
|
||||
|
||||
Instead, report privately by emailing:
|
||||
- `kieran@every.to`
|
||||
|
||||
Include:
|
||||
- A clear description of the issue
|
||||
- Reproduction steps or proof of concept
|
||||
- Impact assessment (what an attacker can do)
|
||||
- Any suggested mitigation
|
||||
|
||||
We will acknowledge receipt as soon as possible and work with you on validation, remediation, and coordinated disclosure timing.
|
||||
|
||||
## Scope Notes
|
||||
|
||||
This repository primarily contains plugin instructions/configuration plus a conversion/install CLI.
|
||||
|
||||
- Plugin instruction content itself does not run as a server process.
|
||||
- Security/privacy behavior also depends on the host AI tool and any external integrations you explicitly invoke.
|
||||
|
||||
For data-handling details, see [PRIVACY.md](PRIVACY.md).
|
||||
@@ -0,0 +1,117 @@
|
||||
---
|
||||
date: 2026-02-14
|
||||
topic: copilot-converter-target
|
||||
---
|
||||
|
||||
# Add GitHub Copilot Converter Target
|
||||
|
||||
## What We're Building
|
||||
|
||||
A new converter target that transforms the compound-engineering Claude Code plugin into GitHub Copilot's native format. This follows the same established pattern as the existing converters (Cursor, Codex, OpenCode, Droid, Pi) and outputs files that Copilot can consume directly from `.github/` (repo-level) or `~/.copilot/` (user-wide).
|
||||
|
||||
Copilot's customization system (as of early 2026) supports: custom agents (`.agent.md`), agent skills (`SKILL.md`), prompt files (`.prompt.md`), custom instructions (`copilot-instructions.md`), and MCP servers (via repo settings).
|
||||
|
||||
## Why This Approach
|
||||
|
||||
The repository already has a robust multi-target converter infrastructure with a consistent `TargetHandler` pattern. Adding Copilot as a new target follows this proven pattern rather than inventing something new. Copilot's format is close enough to Claude Code's that the conversion is straightforward, and the SKILL.md format is already cross-compatible.
|
||||
|
||||
### Approaches Considered
|
||||
|
||||
1. **Full converter target (chosen)** — Follow the existing pattern with types, converter, writer, and target registration. Most consistent with codebase conventions.
|
||||
2. **Minimal agent-only converter** — Only convert agents, skip commands/skills. Too limited; users would lose most of the plugin's value.
|
||||
3. **Documentation-only approach** — Just document how to manually set up Copilot. Doesn't compound — every user would repeat the work.
|
||||
|
||||
## Key Decisions
|
||||
|
||||
### Component Mapping
|
||||
|
||||
| Claude Code Component | Copilot Equivalent | Notes |
|
||||
|----------------------|-------------------|-------|
|
||||
| **Agents** (`.md`) | **Custom Agents** (`.agent.md`) | Full frontmatter mapping: description, tools, target, infer |
|
||||
| **Commands** (`.md`) | **Agent Skills** (`SKILL.md`) | Commands become skills since Copilot has no direct command equivalent. `allowed-tools` dropped silently. |
|
||||
| **Skills** (`SKILL.md`) | **Agent Skills** (`SKILL.md`) | Copy as-is — format is already cross-compatible |
|
||||
| **MCP Servers** | **Repo settings JSON** | Generate a `copilot-mcp-config.json` users paste into GitHub repo settings |
|
||||
| **Hooks** | **Skipped with warning** | Copilot doesn't have a hooks equivalent |
|
||||
|
||||
### Agent Frontmatter Mapping
|
||||
|
||||
| Claude Field | Copilot Field | Mapping |
|
||||
|-------------|--------------|---------|
|
||||
| `name` | `name` | Direct pass-through |
|
||||
| `description` | `description` (required) | Direct pass-through, generate fallback if missing |
|
||||
| `capabilities` | Body text | Fold into body as "## Capabilities" section (like Cursor) |
|
||||
| `model` | `model` | Pass through (works in IDE, may be ignored on github.com) |
|
||||
| — | `tools` | Default to `["*"]` (all tools). Claude agents have unrestricted tool access, so Copilot agents should too. |
|
||||
| — | `target` | Omit (defaults to `both` — IDE + github.com) |
|
||||
| — | `infer` | Set to `true` (auto-selection enabled) |
|
||||
|
||||
### Output Directories
|
||||
|
||||
- **Repository-level (default):** `.github/agents/`, `.github/skills/`
|
||||
- **User-wide (with --personal flag):** `~/.copilot/skills/` (only skills supported at this level)
|
||||
|
||||
### Content Transformation
|
||||
|
||||
Apply transformations similar to Cursor converter:
|
||||
|
||||
1. **Task agent calls:** `Task agent-name(args)` → `Use the agent-name skill to: args`
|
||||
2. **Slash commands:** `/workflows:plan` → `/plan` (flatten namespace)
|
||||
3. **Path rewriting:** `.claude/` → `.github/` (Copilot's repo-level config path)
|
||||
4. **Agent references:** `@agent-name` → `the agent-name agent`
|
||||
|
||||
### MCP Server Handling
|
||||
|
||||
Generate a `copilot-mcp-config.json` file with the structure Copilot expects:
|
||||
|
||||
```json
|
||||
{
|
||||
"mcpServers": {
|
||||
"server-name": {
|
||||
"type": "local",
|
||||
"command": "npx",
|
||||
"args": ["package"],
|
||||
"tools": ["*"],
|
||||
"env": {
|
||||
"KEY": "COPILOT_MCP_KEY"
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
Note: Copilot requires env vars to use the `COPILOT_MCP_` prefix. The converter should transform env var names accordingly and include a comment/note about this.
|
||||
|
||||
## Files to Create/Modify
|
||||
|
||||
### New Files
|
||||
|
||||
- `src/types/copilot.ts` — Type definitions (CopilotAgent, CopilotSkill, CopilotBundle, etc.)
|
||||
- `src/converters/claude-to-copilot.ts` — Converter with `transformContentForCopilot()`
|
||||
- `src/targets/copilot.ts` — Writer with `writeCopilotBundle()`
|
||||
- `docs/specs/copilot.md` — Format specification document
|
||||
|
||||
### Modified Files
|
||||
|
||||
- `src/targets/index.ts` — Register copilot target handler
|
||||
- `src/commands/sync.ts` — Add "copilot" to valid sync targets
|
||||
|
||||
### Test Files
|
||||
|
||||
- `tests/copilot-converter.test.ts` — Converter tests following existing patterns
|
||||
|
||||
### Character Limit
|
||||
|
||||
Copilot imposes a 30,000 character limit on agent body content. If an agent body exceeds this after folding in capabilities, the converter should truncate with a warning to stderr.
|
||||
|
||||
### Agent File Extension
|
||||
|
||||
Use `.agent.md` (not plain `.md`). This is the canonical Copilot convention and makes agent files immediately identifiable.
|
||||
|
||||
## Open Questions
|
||||
|
||||
- Should the converter generate a `copilot-setup-steps.yml` workflow file for MCP servers that need special dependencies (e.g., `uv`, `pipx`)?
|
||||
- Should `.github/copilot-instructions.md` be generated with any base instructions from the plugin?
|
||||
|
||||
## Next Steps
|
||||
|
||||
→ `/workflows:plan` for implementation details
|
||||
@@ -0,0 +1,30 @@
|
||||
---
|
||||
date: 2026-02-17
|
||||
topic: copilot-skill-naming
|
||||
---
|
||||
|
||||
# Copilot Skill Naming: Preserve Namespace
|
||||
|
||||
## What We're Building
|
||||
|
||||
Change the Copilot converter to preserve command namespaces when converting commands to skills. Currently `workflows:plan` flattens to `plan`, which is too generic and clashes with Copilot's own features in the chat suggestion UI.
|
||||
|
||||
## Why This Approach
|
||||
|
||||
The `flattenCommandName` function strips everything before the last colon, producing names like `plan`, `review`, `work` that are too generic for Copilot's skill discovery UI. Replacing colons with hyphens (`workflows:plan` -> `workflows-plan`) preserves context while staying within valid filename characters.
|
||||
|
||||
## Key Decisions
|
||||
|
||||
- **Replace colons with hyphens** instead of stripping the prefix: `workflows:plan` -> `workflows-plan`
|
||||
- **Copilot only** — other converters (Cursor, Droid, etc.) keep their current flattening behavior
|
||||
- **Content transformation too** — slash command references in body text also use hyphens: `/workflows:plan` -> `/workflows-plan`
|
||||
|
||||
## Changes Required
|
||||
|
||||
1. `src/converters/claude-to-copilot.ts` — change `flattenCommandName` to replace colons with hyphens
|
||||
2. `src/converters/claude-to-copilot.ts` — update `transformContentForCopilot` slash command rewriting
|
||||
3. `tests/copilot-converter.test.ts` — update affected tests
|
||||
|
||||
## Next Steps
|
||||
|
||||
-> Implement directly (small, well-scoped change)
|
||||
85
docs/brainstorms/2026-03-14-ce-plan-rewrite-requirements.md
Normal file
85
docs/brainstorms/2026-03-14-ce-plan-rewrite-requirements.md
Normal file
@@ -0,0 +1,85 @@
|
||||
---
|
||||
date: 2026-03-14
|
||||
topic: ce-plan-rewrite
|
||||
---
|
||||
|
||||
# Rewrite `ce:plan` to Separate Planning from Implementation
|
||||
|
||||
## Problem Frame
|
||||
|
||||
`ce:plan` sits between `ce:brainstorm` and `ce:work`, but the current skill mixes issue authoring, technical planning, and pseudo-implementation. That makes plans brittle and pushes the planning phase to predict details that are often only discoverable during implementation. PR #246 intensifies this by asking plans to include complete code, exact commands, and micro-step TDD and commit choreography. The rewrite should keep planning strong enough for a capable agent or engineer to execute, while moving code-writing, test-running, and execution-time learning back into `ce:work`.
|
||||
|
||||
## Requirements
|
||||
|
||||
- R1. `ce:plan` must accept either a raw feature description or a requirements document produced by `ce:brainstorm` as primary input.
|
||||
- R2. `ce:plan` must preserve compound-engineering's planning strengths: repo pattern scan, institutional learnings, conditional external research, and requirements-gap checks when warranted.
|
||||
- R3. `ce:plan` must produce a durable implementation plan focused on decisions, sequencing, file paths, dependencies, risks, and test scenarios, not implementation code.
|
||||
- R4. `ce:plan` must not instruct the planner to run tests, generate exact implementation snippets, or learn from execution-time results. Those belong to `ce:work`.
|
||||
- R5. Plan tasks and subtasks must be right-sized for implementation handoff, but sized as logical units or atomic commits rather than 2-5 minute copy-paste steps.
|
||||
- R6. Plans must remain shareable and portable as documents or issues without tool-specific executor litter such as TodoWrite instructions, `/ce:work` choreography, or git command recipes in the artifact itself.
|
||||
- R7. `ce:plan` must carry forward product decisions, scope boundaries, success criteria, and deferred questions from `ce:brainstorm` without re-inventing them.
|
||||
- R8. `ce:plan` must explicitly distinguish what gets resolved during planning from what is intentionally deferred to implementation-time discovery.
|
||||
- R9. `ce:plan` must hand off cleanly to `ce:work`, giving enough information for task creation without pre-writing code.
|
||||
- R10. If detail levels remain, they must change depth of analysis and documentation, not the planning philosophy. A small plan can be terse while still staying decision-first.
|
||||
- R11. If an upstream requirements document contains unresolved `Resolve Before Planning` items, `ce:plan` must classify whether they are true product blockers or misfiled technical questions before proceeding.
|
||||
- R12. `ce:plan` must not plan past unresolved product decisions that would change behavior, scope, or success criteria, but it may absorb technical or research questions by reclassifying them into planning-owned investigation.
|
||||
- R13. When true blockers remain, `ce:plan` must pause helpfully: surface the blockers, allow the user to convert them into explicit assumptions or decisions, or route them back to `ce:brainstorm`.
|
||||
|
||||
## Success Criteria
|
||||
|
||||
- A fresh implementer can start work from the plan without needing clarifying questions, but the plan does not contain implementation code.
|
||||
- `ce:work` can derive actionable tasks from the plan without relying on micro-step commands or embedded git/test instructions.
|
||||
- Plans stay accurate longer as repo context changes because they capture decisions and boundaries rather than speculative code.
|
||||
- A requirements document from `ce:brainstorm` flows into planning without losing decisions, scope boundaries, or success criteria.
|
||||
- Plans do not proceed past unresolved product blockers unless the user explicitly converts them into assumptions or decisions.
|
||||
- For the same feature, the rewritten `ce:plan` produces output that is materially shorter and less brittle than the current skill or PR #246's proposed format while remaining execution-ready.
|
||||
|
||||
## Scope Boundaries
|
||||
|
||||
- Do not redesign `ce:brainstorm`'s product-definition role.
|
||||
- Do not remove decomposition, file paths, verification, or risk analysis from `ce:plan`.
|
||||
- Do not move planning into a vague, under-specified artifact that leaves execution to guess.
|
||||
- Do not change `ce:work` in this phase beyond possible follow-up clarification of what plan structure it should prefer.
|
||||
- Do not require heavyweight PRD ceremony for small or straightforward work.
|
||||
|
||||
## Key Decisions
|
||||
|
||||
- Use a hybrid model: keep compound-engineering's research and handoff strengths, but adopt iterative-engineering's "decisions, not code" boundary.
|
||||
- Planning stops before execution: no running tests, no fail/pass learning, no exact implementation snippets, and no commit shell commands in the plan.
|
||||
- Use logical tasks and subtasks sized around atomic changes or commit units rather than 2-5 minute micro-steps.
|
||||
- Keep explicit verification and test scenarios, but express them as expected coverage and validation outcomes rather than commands with predicted output.
|
||||
- Preserve `ce:brainstorm` as the preferred upstream input when available, with clear handling for deferred technical questions.
|
||||
- Treat `Resolve Before Planning` as a classification gate: planning first distinguishes true product blockers from technical questions, then investigates only the latter.
|
||||
|
||||
## High-Level Direction
|
||||
|
||||
- Phase 0: Resume existing plan work when relevant, detect brainstorm input, and assess scope.
|
||||
- Phase 1: Gather context through repo research, institutional learnings, and conditional external research.
|
||||
- Phase 2: Resolve planning-time technical questions and capture implementation-time unknowns separately.
|
||||
- Phase 3: Structure the plan around components, dependencies, files, test targets, risks, and verification.
|
||||
- Phase 4: Write a right-sized plan artifact whose depth varies by scope, but whose boundary stays planning-only.
|
||||
- Phase 5: Review and hand off to refinement, deeper research, issue sharing, or `ce:work`.
|
||||
|
||||
## Alternatives Considered
|
||||
|
||||
- Keep the current `ce:plan` and only reject PR #246.
|
||||
Rejected because the underlying issue remains: the current skill already drifts toward issue-template output plus pseudo-implementation.
|
||||
- Adopt Superpowers `writing-plans` nearly wholesale.
|
||||
Rejected because it is intentionally execution-script-oriented and collapses planning into detailed code-writing and command choreography.
|
||||
- Adopt iterative-engineering `tech-planning` wholesale.
|
||||
Rejected because it would lose useful compound-engineering behaviors such as brainstorm-origin integration, institutional learnings, and richer post-plan handoff options.
|
||||
|
||||
## Dependencies / Assumptions
|
||||
|
||||
- `ce:work` can continue creating its own actionable task list from a decision-first plan.
|
||||
- If `ce:work` later benefits from an explicit section such as `## Implementation Units` or `## Work Breakdown`, that should be a separate follow-up designed around execution needs rather than micro-step code generation.
|
||||
|
||||
## Resolved During Planning
|
||||
|
||||
- [Affects R10][Technical] Replaced `MINIMAL` / `MORE` / `A LOT` with `Lightweight` / `Standard` / `Deep` to align `ce:plan` with `ce:brainstorm`'s scope model.
|
||||
- [Affects R9][Technical] Updated `ce:work` to explicitly consume decision-first plan sections such as `Implementation Units`, `Requirements Trace`, `Files`, `Test Scenarios`, and `Verification`.
|
||||
- [Affects R2][Needs research] Kept SpecFlow as a conditional planning aid: use it for `Standard` or `Deep` plans when flow completeness is unclear rather than making it mandatory for every plan.
|
||||
|
||||
## Next Steps
|
||||
|
||||
-> Review, refine, and commit the `ce:plan` and `ce:work` rewrite
|
||||
77
docs/brainstorms/2026-03-15-ce-ideate-skill-requirements.md
Normal file
77
docs/brainstorms/2026-03-15-ce-ideate-skill-requirements.md
Normal file
@@ -0,0 +1,77 @@
|
||||
---
|
||||
date: 2026-03-15
|
||||
topic: ce-ideate-skill
|
||||
---
|
||||
|
||||
# ce:ideate — Open-Ended Ideation Skill
|
||||
|
||||
## Problem Frame
|
||||
|
||||
The ce:brainstorm skill is reactive — the user brings an idea, and the skill helps refine it through collaborative dialogue. There is no workflow for the opposite direction: having the AI proactively generate ideas by deeply understanding the project and then filtering them through critical self-evaluation. Users currently achieve this through ad-hoc prompting (e.g., "come up with 100 ideas and give me your best 10"), but that approach has no codebase grounding, no structured output, no durable artifact, and no connection to the ce:* workflow pipeline.
|
||||
|
||||
## Requirements
|
||||
|
||||
- R1. ce:ideate is a standalone skill, separate from ce:brainstorm, with its own SKILL.md in `plugins/compound-engineering/skills/ce-ideate/`
|
||||
- R2. Accepts an optional freeform argument that serves as a focus hint — can be a concept ("DX improvements"), a path ("plugins/compound-engineering/skills/"), a constraint ("low-complexity quick wins"), or empty for fully open ideation
|
||||
- R3. Performs a deep codebase scan before generating ideas, grounding ideation in the actual project state rather than abstract speculation
|
||||
- R4. Preserves the user's proven prompt mechanism as the core workflow: generate many ideas first, then systematically and critically reject weak ones, then explain only the surviving ideas in detail
|
||||
- R5. Self-critiques the full list, rejecting weak ideas with explicit reasoning — the adversarial filtering step is the core quality mechanism
|
||||
- R6. Presents the top 5-7 surviving ideas with structured analysis: description, rationale, downsides, confidence score (0-100%), estimated complexity
|
||||
- R7. Includes a brief rejection summary — one-line per rejected idea with the reason — so the user can see what was considered and why it was cut
|
||||
- R8. Writes a durable ideation artifact to `docs/ideation/YYYY-MM-DD-<topic>-ideation.md` (or `YYYY-MM-DD-open-ideation.md` when no focus area). This compounds — rejected ideas prevent re-exploring dead ends, and un-acted-on ideas remain available for future sessions.
|
||||
- R9. The default volume (~30 ideas, top 5-7 presented) can be overridden by the user's argument (e.g., "give me your top 3" or "go deep, 100 ideas")
|
||||
- R10. Handoff options after presenting ideas: brainstorm a selected idea (feeds into ce:brainstorm), refine the ideation (dig deeper, re-evaluate, explore new angles), share to Proof, or end the session
|
||||
- R11. Always routes to ce:brainstorm when the user wants to act on an idea — ideation output is never detailed enough to skip requirements refinement
|
||||
- R12. Session completion: when ending, offer to commit the ideation doc to the current branch. If the user declines, leave the file uncommitted. Do not create branches or push — just the local commit.
|
||||
- R13. Resume behavior: when ce:ideate is invoked, check `docs/ideation/` for ideation docs created within the last 30 days. If a relevant one exists, offer to continue from it (add new ideas, revisit rejected ones, act on un-explored ideas) or start fresh.
|
||||
- R14. Present the surviving candidates to the user before writing the durable ideation artifact, so the user can ask questions or lightly reshape the candidate set before it is archived
|
||||
- R15. The ideation artifact must be written or updated before any downstream handoff, Proof sharing, or session end, even though the initial survivor presentation happens first
|
||||
- R16. Refine routes based on intent: "add more ideas" or "explore new angles" returns to generation (Phase 2), "re-evaluate" or "raise the bar" returns to critique (Phase 3), "dig deeper on idea #N" expands that idea's analysis in place. The ideation doc is updated after each refinement when the refined state is being preserved
|
||||
- R17. Uses agent intelligence to improve ideation quality, but only as support for the core prompt mechanism rather than as a replacement for it
|
||||
- R18. Uses existing research agents for codebase grounding, but ideation and critique sub-agents are prompt-defined roles with distinct perspectives rather than forced reuse of existing named review agents
|
||||
- R19. When sub-agents are used for ideation, each one receives the same grounding summary, the user focus hint, and the current volume target
|
||||
- R20. Focus hints influence both candidate generation and final filtering; they are not only an evaluation-time bias
|
||||
- R21. Ideation sub-agents return ideas in a standardized structured format so the orchestrator can merge, dedupe, and reason over them consistently
|
||||
- R22. The orchestrator owns final scoring, ranking, and survivor decisions across the merged idea set; sub-agents may emit lightweight local signals, but they do not authoritatively rank their own ideas
|
||||
- R23. Distinct ideation perspectives should be created through prompt framing methods that encourage creative spread without over-constraining the workflow; examples include friction, unmet need, inversion, assumption-breaking, leverage, and extreme-case prompts
|
||||
- R24. The skill does not hardcode a fixed number of sub-agents for all runs; it should use the smallest useful set that preserves diversity without overwhelming the orchestrator's context window
|
||||
- R25. When the user picks an idea to brainstorm, the ideation doc is updated to mark that idea as "explored" with a reference to the resulting brainstorm session date, so future revisits show which ideas have been acted on.
|
||||
|
||||
## Success Criteria
|
||||
|
||||
- A user can invoke `/ce:ideate` with no arguments on any project and receive genuinely surprising, high-quality improvement ideas grounded in the actual codebase
|
||||
- Ideas that survive the filter are meaningfully better than what the user would get from a naive "give me 10 ideas" prompt
|
||||
- The workflow uses agent intelligence to widen the candidate pool without obscuring the core generate -> reject -> survivors mechanism
|
||||
- The user sees and can question the surviving candidates before they are written into the durable artifact
|
||||
- The ideation artifact persists and provides value when revisited weeks later
|
||||
- The skill composes naturally with the existing pipeline: ideate → brainstorm → plan → work
|
||||
|
||||
## Scope Boundaries
|
||||
|
||||
- ce:ideate does NOT produce requirements, plans, or code — it produces ranked ideas
|
||||
- ce:ideate does NOT modify ce:brainstorm's behavior — discovery of ce:ideate is handled through the skill description and catalog, not by altering other skills
|
||||
- The skill does not do external research (competitive analysis, similar projects) in v1 — this could be a future enhancement but adds cost and latency without proven need
|
||||
- No configurable depth modes in v1 — fixed volume with argument-based override is sufficient
|
||||
|
||||
## Key Decisions
|
||||
|
||||
- **Standalone skill, not a mode within ce:brainstorm**: The workflows are fundamentally different cognitive modes (proactive/divergent vs. reactive/convergent) with different phases, outputs, and success criteria. Combining them would make ce:brainstorm harder to maintain and blur its identity.
|
||||
- **Durable artifact in docs/ideation/**: Discarding ideation results is anti-compounding. The file is cheap to write and provides value when revisiting un-acted-on ideas or avoiding re-exploration of rejected ones.
|
||||
- **Artifact written after candidate review, not before initial presentation**: The first survivor presentation is collaborative review, not archival finalization. The artifact should be written only after the candidate set is good enough to preserve, but always before handoff, sharing, or session end.
|
||||
- **Always route to ce:brainstorm for follow-up**: At ideation depth, ideas are one-paragraph concepts — never detailed enough to skip requirements refinement.
|
||||
- **Survivors + rejection summary output format**: Full transparency on what was considered without overwhelming with detailed analysis of rejected ideas.
|
||||
- **Freeform optional argument**: A concept, a path, or nothing at all — the skill interprets whatever it gets as context. No artificial distinction between "focus area" and "target path."
|
||||
- **Agent intelligence as support, not replacement**: The value comes from the proven ideation-and-rejection mechanism. Parallel sub-agents help produce a richer candidate pool and stronger critique, but the orchestrator remains responsible for synthesis, scoring, and final ranking.
|
||||
|
||||
## Outstanding Questions
|
||||
|
||||
### Deferred to Planning
|
||||
|
||||
- [Affects R3][Technical] Which research agents should always run for codebase grounding in v1 beyond `repo-research-analyst` and `learnings-researcher`, if any?
|
||||
- [Affects R21][Technical] What exact structured output schema should ideation sub-agents return so the orchestrator can merge and score consistently without overfitting the format too early?
|
||||
- [Affects R6][Technical] Should the structured analysis per surviving idea include "suggested next steps" or "what this would unlock" beyond the current fields (description, rationale, downsides, confidence, complexity)?
|
||||
- [Affects R2][Technical] How should the skill detect volume overrides in the freeform argument vs. focus-area hints? Simple heuristic or explicit parsing?
|
||||
|
||||
## Next Steps
|
||||
|
||||
→ `/ce:plan` for structured implementation planning
|
||||
@@ -0,0 +1,65 @@
|
||||
---
|
||||
date: 2026-03-16
|
||||
topic: issue-grounded-ideation
|
||||
---
|
||||
|
||||
# Issue-Grounded Ideation Mode for ce:ideate
|
||||
|
||||
## Problem Frame
|
||||
|
||||
When a team wants to ideate on improvements, their issue tracker holds rich signal about real user pain, recurring failures, and severity patterns — but ce:ideate currently only looks at the codebase and past learnings. Teams have to manually synthesize issue patterns before ideating, or they ideate without that context and miss what their users are actually hitting.
|
||||
|
||||
The goal is not "fix individual bugs" but "generate strategic improvement ideas grounded in the patterns your issue tracker reveals." 25 duplicate bugs about the same failure mode is a signal about collaboration reliability, not 25 separate problems.
|
||||
|
||||
## Requirements
|
||||
|
||||
- R1. When the user's argument indicates they want issue-tracker data as input (e.g., "bugs", "github issues", "open issues", "what users are reporting", "issue patterns"), ce:ideate activates an issue intelligence step alongside the existing Phase 1 scans
|
||||
- R2. A new **issue intelligence agent** fetches, clusters, deduplicates, and analyzes issues, returning structured theme analysis — not a list of individual issues
|
||||
- R3. The agent fetches **open issues** plus **recently closed issues** (approximately 30 days), filtering out issues closed as duplicate, won't-fix, or not-planned. Recently fixed issues are included because they show which areas had enough pain to warrant action.
|
||||
- R4. Issue clusters drive the ideation frames in Phase 2 using a **hybrid strategy**: derive frames from clusters, pad with default frames (e.g., "assumption-breaking", "leverage/compounding") when fewer than 4 clusters exist. This ensures ideas are grounded in real pain patterns while maintaining ideation diversity.
|
||||
- R5. The existing Phase 1 scans (codebase context + learnings search) still run in parallel — issue analysis is additive context, not a replacement
|
||||
- R6. The issue intelligence agent detects the repository from the current directory's git remote
|
||||
- R7. Start with GitHub issues via `gh` CLI. Design the agent prompt and output structure so Linear or other trackers can be added later without restructuring the ideation flow.
|
||||
- R8. The issue intelligence agent is independently useful outside of ce:ideate — it can be dispatched directly by a user or other workflows to summarize issue themes, understand the current landscape, or reason over recent activity. Its output should be self-contained, not coupled to ideation-specific context.
|
||||
- R9. The agent's output must communicate at the **theme level**, not the individual-issue level. Each theme should convey: what the pattern is, why it matters (user impact, severity, frequency, trend direction), and what it signals about the system. The output should help a human or agent fully understand the importance and shape of each theme without needing to read individual issues.
|
||||
|
||||
## Success Criteria
|
||||
|
||||
- Running `/ce:ideate bugs` on a repo with noisy/duplicate issues (like proof's 25+ LIVE_DOC_UNAVAILABLE variants) produces clustered themes, not a rehash of individual issues
|
||||
- Surviving ideas are strategic improvements ("invest in collaboration reliability infrastructure") not bug fixes ("fix LIVE_DOC_UNAVAILABLE")
|
||||
- The issue intelligence agent's output is structured enough that ideation sub-agents can engage with themes meaningfully
|
||||
- Ideation quality is at least as good as the default mode, with the added benefit of issue grounding
|
||||
|
||||
## Scope Boundaries
|
||||
|
||||
- GitHub issues only in v1 (Linear is a future extension)
|
||||
- No issue triage or management — this is read-only analysis for ideation input
|
||||
- No changes to Phase 3 (adversarial filtering) or Phase 4 (presentation) — only Phase 1 and Phase 2 frame derivation are affected
|
||||
- The issue intelligence agent is a new agent file, not a modification to an existing research agent
|
||||
- The agent is designed as a standalone capability that ce:ideate composes, not an ideation-internal module
|
||||
- Assumes `gh` CLI is available and authenticated in the environment
|
||||
- When a repo has too few issues to cluster meaningfully (e.g., < 5 open+recent), the agent should report that and ce:ideate should fall back to default ideation with a note to the user
|
||||
|
||||
## Key Decisions
|
||||
|
||||
- **Pattern-first, not issue-first**: The output is improvement ideas grounded in bug patterns, not a prioritized bug list. The ideation instructions already prevent "just fix bug #534" thinking.
|
||||
- **Hybrid frame strategy**: Clusters derive ideation frames, padded with defaults when thin. Pure cluster-derived frames risk too few frames; pure default frames risk ignoring the issue signal.
|
||||
- **Flexible argument detection**: Use intent-based parsing ("reasonable interpretation rather than formal parsing") consistent with the existing volume hint system. No rigid keyword matching.
|
||||
- **Open + recently closed**: Including recently fixed issues provides richer pattern data — shows which areas warranted action, not just what's currently broken.
|
||||
- **Additive to Phase 1**: Issue analysis runs as a third parallel agent alongside codebase scan and learnings search. All three feed the grounding summary.
|
||||
- **Titles + labels + sample bodies**: Read titles and labels for all issues (cheap), then read full bodies for 2-3 representative issues per emerging cluster. This handles both well-labeled repos (labels drive clustering, bodies confirm) and poorly-labeled repos (bodies drive clustering). Avoids reading all bodies which is expensive at scale.
|
||||
|
||||
## Outstanding Questions
|
||||
|
||||
### Deferred to Planning
|
||||
|
||||
- [Affects R2][Technical] What structured output format should the issue intelligence agent return? Likely theme clusters with: theme name, issue count, severity distribution, representative issue titles, and a one-line synthesis.
|
||||
- [Affects R3][Technical] How to detect GitHub close reasons (completed vs not-planned vs duplicate) via `gh` CLI? May need `gh issue list --state closed --json stateReason` or label-based filtering.
|
||||
- [Affects R4][Technical] What's the threshold for "too few clusters"? Current thinking: pad with default frames when fewer than 4 clusters, but this may need tuning.
|
||||
- [Affects R6][Technical] How to extract the GitHub repo from git remote? Standard `gh repo view --json nameWithOwner` or parse the remote URL.
|
||||
- [Affects R7][Needs research] What would a Linear integration look like? Just swapping the fetch mechanism, or does Linear's project/cycle structure change the clustering approach?
|
||||
- [Affects R2][Technical] Exact number of sample bodies per cluster to read (starting point: 2-3 per cluster).
|
||||
|
||||
## Next Steps
|
||||
|
||||
→ `/ce:plan` for structured implementation planning
|
||||
@@ -0,0 +1,89 @@
|
||||
---
|
||||
date: 2026-03-17
|
||||
topic: release-automation
|
||||
---
|
||||
|
||||
# Release Automation and Changelog Ownership
|
||||
|
||||
## Problem Frame
|
||||
|
||||
The repository currently has one automated release flow for the npm CLI, but the broader release story is split across CI, manual maintainer workflows, stale docs, and multiple version surfaces. That makes it hard to batch releases intentionally, hard for multiple maintainers to share release responsibility, and easy for changelogs, plugin manifests, and derived metadata like component counts to drift out of sync. The goal is to move to a release model that supports intentional batching, independent component versioning, centralized history, and CI-owned release authority without forcing version bumps for untouched plugins.
|
||||
|
||||
## Requirements
|
||||
|
||||
- R1. The release process must be manually triggered; merging to `main` must not automatically publish a release.
|
||||
- R2. The release system must support batching: releasable merges may accumulate on `main` until maintainers decide to cut a release.
|
||||
- R3. The release system must maintain a single release PR for the whole repo that stays open until merged and automatically accumulates additional releasable changes merged to `main`.
|
||||
- R4. The release system must support independent version bumps for these components: `cli`, `compound-engineering`, `coding-tutor`, and `marketplace`.
|
||||
- R5. The release system must not bump untouched plugins or unrelated components.
|
||||
- R6. The release system must preserve one centralized root `CHANGELOG.md` as the canonical changelog for the repository.
|
||||
- R7. The root changelog must record releases as top-level entries per component version, rather than requiring separate changelog files per plugin.
|
||||
- R8. Existing root changelog history must be preserved during the migration; the new release model must not discard or rewrite historical entries in a way that loses continuity.
|
||||
- R9. `plugins/compound-engineering/CHANGELOG.md` must no longer be treated as the canonical changelog after the migration.
|
||||
- R10. The release process must replace the current `release-docs` workflow; `release-docs` must no longer act as a release authority or required release step.
|
||||
- R11. Narrow scripts must replace `release-docs` responsibilities, including metadata synchronization, count calculation, docs generation where still needed, and validation.
|
||||
- R12. Release automation must be the sole authority for version bumps, changelog writes, and computed metadata updates such as counts of agents, skills, commands, or similar release-owned descriptions.
|
||||
- R13. The release flow must support a dry-run mode that summarizes what would happen without publishing, tagging, or committing release changes.
|
||||
- R14. Dry run output must clearly summarize which components would release, the proposed version bumps, the changelog entries that would be added, and any blocking validation failures.
|
||||
- R15. Marketplace version bumps must happen only for marketplace-level changes, such as marketplace metadata changes or adding/removing plugins from the catalog.
|
||||
- R16. Updating a plugin version alone must not require a marketplace version bump.
|
||||
- R17. Plugin-only content changes must be releasable without requiring a CLI version bump when the CLI code itself has not changed.
|
||||
- R18. The release model must remain compatible with the current install behavior where `bunx @every-env/compound-plugin install ...` runs the npm CLI but fetches named plugin content from the GitHub repository at runtime.
|
||||
- R19. The release process must be triggerable by a maintainer or an AI agent through CI without requiring a local maintainer-only skill.
|
||||
- R20. The resulting model must scale to future plugins without requiring the repo to special-case `compound-engineering` forever.
|
||||
- R21. The release model must continue to rely on conventional release intent signals (`feat`, `fix`, breaking changes, etc.), but component scopes in commit or PR titles must remain optional rather than required.
|
||||
- R22. Release automation must infer component ownership primarily from changed files, not from commit or PR title scopes alone.
|
||||
- R23. The repo should enforce parseable conventional PR or merge titles strongly enough for release tooling to classify change type, while avoiding mandatory component scoping on every change.
|
||||
- R24. The manual CI-driven release workflow must support explicit bump overrides for exceptional cases, at least `patch`, `minor`, and `major`, without requiring maintainers to create fake or empty commits purely to coerce a release.
|
||||
- R25. Bump overrides must be expressible per component rather than only as a repo-wide override.
|
||||
- R26. Dry run output must clearly show both the inferred bump and any applied manual override for each affected component.
|
||||
|
||||
## Success Criteria
|
||||
|
||||
- Maintainers can let multiple PRs merge to `main` without immediately cutting a release.
|
||||
- At any point, maintainers can inspect a release PR or dry run and understand what would ship next.
|
||||
- A change to `coding-tutor` does not force a version bump to `compound-engineering`.
|
||||
- A plugin version bump does not force a marketplace version bump unless marketplace-level files changed.
|
||||
- Release-owned metadata and counts stay in sync without relying on a local slash command.
|
||||
- The root changelog remains readable and continuous before and after the migration.
|
||||
|
||||
## Scope Boundaries
|
||||
|
||||
- This work does not require changing how Claude Code itself consumes plugin and marketplace versions.
|
||||
- This work does not require solving end-user auto-update discovery for non-Claude harnesses in v1.
|
||||
- This work does not require adding dedicated per-plugin changelog files as the canonical history model.
|
||||
- This work does not require immediate future automation of release timing; manual release remains the default.
|
||||
|
||||
## Key Decisions
|
||||
|
||||
- **Use `release-please` rather than a single release-line flow**: The repo now has multiple independently versioned components, and the release PR model matches the need to batch merges on `main` until a release is intentionally cut.
|
||||
- **One release PR for the whole repo**: Centralized release visibility matters more than separate PRs per component, and a single release PR can still carry multiple component bumps.
|
||||
- **Manual release timing**: The release process should prepare and accumulate the next release automatically, but the decision to cut that release should remain explicit.
|
||||
- **Root changelog stays canonical**: Centralized history is more important than per-plugin changelog isolation for the current repo shape.
|
||||
- **Top-level changelog entries per component version**: This preserves one changelog file while keeping independent component version history readable.
|
||||
- **Retire `release-docs`**: Its responsibilities are too broad, stale, and conflated. Release logic, docs logic, and metadata synchronization should be separated.
|
||||
- **Scripts for narrow responsibilities**: Explicit scripts are easier to validate, automate, and reuse from CI than a local repo-maintenance skill.
|
||||
- **Marketplace version is catalog-scoped**: Plugin version bumps alone should not imply a marketplace release.
|
||||
- **Conventional type required, component scope optional**: Release intent should still come from conventional commit semantics, but requiring `(compound-engineering)` on most repo changes would add unnecessary wording overhead. Component detection should remain file-driven.
|
||||
- **Manual bump override is an explicit escape hatch**: Automatic bump inference remains the default, but maintainers should be able to override a component's release level in CI for exceptional cases without awkward synthetic commits.
|
||||
|
||||
## Dependencies / Assumptions
|
||||
|
||||
- The current install flow for named plugins continues to fetch plugin content from GitHub at runtime, so plugin content releases can remain independent from CLI releases unless CLI behavior also changes.
|
||||
- Claude Code already respects marketplace and plugin versions, so those version surfaces remain meaningful release signals.
|
||||
|
||||
## Outstanding Questions
|
||||
|
||||
### Deferred to Planning
|
||||
|
||||
- [Affects R3][Technical] Should the release PR be updated automatically on every push to `main`, or via a manually triggered maintenance workflow that refreshes the release PR state on demand?
|
||||
- [Affects R7][Technical] What exact root changelog format best balances readability and automation for multiple component-version entries in one file?
|
||||
- [Affects R11][Technical] Which responsibilities should become distinct scripts versus steps embedded directly in the CI workflow?
|
||||
- [Affects R12][Technical] Which release-owned metadata fields should be computed automatically versus validated and left untouched when no count change is needed?
|
||||
- [Affects R9][Technical] Should `plugins/compound-engineering/CHANGELOG.md` be deleted, frozen, or replaced with a short pointer note after the migration?
|
||||
- [Affects R21][Technical] Should conventional-format enforcement happen on PR titles, squash-merge titles, commits, or some combination of them?
|
||||
- [Affects R24][Technical] Should manual bump overrides be implemented as workflow inputs that shape the generated release PR directly, or as an internal generated release-control commit on the release branch only?
|
||||
|
||||
## Next Steps
|
||||
|
||||
→ `/ce:plan` for structured implementation planning
|
||||
@@ -1,675 +0,0 @@
|
||||
/* Documentation-specific styles */
|
||||
|
||||
/* ============================================
|
||||
Documentation Layout
|
||||
============================================ */
|
||||
|
||||
.docs-layout {
|
||||
display: grid;
|
||||
grid-template-columns: 1fr;
|
||||
min-height: 100vh;
|
||||
}
|
||||
|
||||
@media (min-width: 1024px) {
|
||||
.docs-layout {
|
||||
grid-template-columns: 280px 1fr;
|
||||
}
|
||||
}
|
||||
|
||||
/* ============================================
|
||||
Sidebar
|
||||
============================================ */
|
||||
|
||||
.docs-sidebar {
|
||||
position: fixed;
|
||||
top: 0;
|
||||
left: -300px;
|
||||
width: 280px;
|
||||
height: 100vh;
|
||||
background-color: var(--color-background);
|
||||
border-right: 1px solid var(--color-border);
|
||||
overflow-y: auto;
|
||||
transition: left 0.3s ease;
|
||||
z-index: 100;
|
||||
}
|
||||
|
||||
.docs-sidebar.open {
|
||||
left: 0;
|
||||
}
|
||||
|
||||
@media (min-width: 1024px) {
|
||||
.docs-sidebar {
|
||||
position: sticky;
|
||||
left: 0;
|
||||
}
|
||||
}
|
||||
|
||||
.sidebar-header {
|
||||
padding: var(--space-l);
|
||||
border-bottom: 1px solid var(--color-border);
|
||||
}
|
||||
|
||||
.sidebar-header .nav-brand {
|
||||
display: flex;
|
||||
align-items: center;
|
||||
gap: var(--space-s);
|
||||
text-decoration: none;
|
||||
color: var(--color-text-primary);
|
||||
font-weight: 600;
|
||||
}
|
||||
|
||||
.sidebar-header .logo-icon {
|
||||
color: var(--color-accent);
|
||||
font-size: var(--font-size-l);
|
||||
}
|
||||
|
||||
.sidebar-header .logo-text {
|
||||
display: inline;
|
||||
}
|
||||
|
||||
.sidebar-nav {
|
||||
padding: var(--space-l);
|
||||
}
|
||||
|
||||
.nav-section {
|
||||
margin-bottom: var(--space-xl);
|
||||
}
|
||||
|
||||
.nav-section h3 {
|
||||
font-size: var(--font-size-xs);
|
||||
font-weight: 600;
|
||||
text-transform: uppercase;
|
||||
letter-spacing: 0.05em;
|
||||
color: var(--color-text-tertiary);
|
||||
margin: 0 0 var(--space-m) 0;
|
||||
}
|
||||
|
||||
.nav-section ul {
|
||||
list-style: none;
|
||||
margin: 0;
|
||||
padding: 0;
|
||||
}
|
||||
|
||||
.nav-section li {
|
||||
margin: 0;
|
||||
}
|
||||
|
||||
.nav-section a {
|
||||
display: block;
|
||||
padding: var(--space-s) var(--space-m);
|
||||
color: var(--color-text-secondary);
|
||||
text-decoration: none;
|
||||
font-size: var(--font-size-s);
|
||||
border-radius: var(--radius-s);
|
||||
transition: all 0.2s ease;
|
||||
}
|
||||
|
||||
.nav-section a:hover {
|
||||
color: var(--color-text-primary);
|
||||
background-color: var(--color-surface);
|
||||
}
|
||||
|
||||
.nav-section a.active {
|
||||
color: var(--color-accent);
|
||||
background-color: var(--color-accent-light);
|
||||
}
|
||||
|
||||
/* ============================================
|
||||
Main Content
|
||||
============================================ */
|
||||
|
||||
.docs-content {
|
||||
padding: var(--space-xl);
|
||||
max-width: 900px;
|
||||
}
|
||||
|
||||
@media (min-width: 1024px) {
|
||||
.docs-content {
|
||||
padding: var(--space-xxl);
|
||||
}
|
||||
}
|
||||
|
||||
.docs-header {
|
||||
display: flex;
|
||||
align-items: center;
|
||||
justify-content: space-between;
|
||||
margin-bottom: var(--space-xl);
|
||||
}
|
||||
|
||||
.breadcrumb {
|
||||
display: flex;
|
||||
align-items: center;
|
||||
gap: var(--space-s);
|
||||
font-size: var(--font-size-s);
|
||||
color: var(--color-text-tertiary);
|
||||
}
|
||||
|
||||
.breadcrumb a {
|
||||
color: var(--color-text-secondary);
|
||||
text-decoration: none;
|
||||
}
|
||||
|
||||
.breadcrumb a:hover {
|
||||
color: var(--color-accent);
|
||||
}
|
||||
|
||||
.mobile-menu-toggle {
|
||||
display: flex;
|
||||
align-items: center;
|
||||
justify-content: center;
|
||||
width: 40px;
|
||||
height: 40px;
|
||||
background: none;
|
||||
border: 1px solid var(--color-border);
|
||||
border-radius: var(--radius-s);
|
||||
color: var(--color-text-secondary);
|
||||
cursor: pointer;
|
||||
}
|
||||
|
||||
@media (min-width: 1024px) {
|
||||
.mobile-menu-toggle {
|
||||
display: none;
|
||||
}
|
||||
}
|
||||
|
||||
/* ============================================
|
||||
Article Styles
|
||||
============================================ */
|
||||
|
||||
.docs-article {
|
||||
line-height: 1.7;
|
||||
}
|
||||
|
||||
.docs-article h1 {
|
||||
font-size: var(--font-size-xl);
|
||||
margin-bottom: var(--space-l);
|
||||
}
|
||||
|
||||
.docs-article h2 {
|
||||
font-size: var(--font-size-l);
|
||||
margin-top: var(--space-xxl);
|
||||
margin-bottom: var(--space-l);
|
||||
padding-bottom: var(--space-s);
|
||||
border-bottom: 1px solid var(--color-border);
|
||||
display: flex;
|
||||
align-items: center;
|
||||
gap: var(--space-s);
|
||||
}
|
||||
|
||||
.docs-article h2 i {
|
||||
color: var(--color-accent);
|
||||
}
|
||||
|
||||
.docs-article h3 {
|
||||
font-size: var(--font-size-m);
|
||||
margin-top: var(--space-xl);
|
||||
margin-bottom: var(--space-m);
|
||||
}
|
||||
|
||||
.docs-article h4 {
|
||||
font-size: var(--font-size-s);
|
||||
margin-top: var(--space-l);
|
||||
margin-bottom: var(--space-s);
|
||||
}
|
||||
|
||||
.docs-article p {
|
||||
margin-bottom: var(--space-l);
|
||||
}
|
||||
|
||||
.docs-article .lead {
|
||||
font-size: var(--font-size-l);
|
||||
color: var(--color-text-secondary);
|
||||
margin-bottom: var(--space-xl);
|
||||
}
|
||||
|
||||
.docs-article ul,
|
||||
.docs-article ol {
|
||||
margin-bottom: var(--space-l);
|
||||
padding-left: var(--space-xl);
|
||||
}
|
||||
|
||||
.docs-article li {
|
||||
margin-bottom: var(--space-s);
|
||||
}
|
||||
|
||||
/* ============================================
|
||||
Code Blocks in Docs
|
||||
============================================ */
|
||||
|
||||
.docs-article .card-code-block {
|
||||
margin: var(--space-l) 0;
|
||||
}
|
||||
|
||||
.docs-article code {
|
||||
font-family: var(--font-mono);
|
||||
font-size: 0.9em;
|
||||
background-color: var(--color-surface);
|
||||
padding: 2px 6px;
|
||||
border-radius: var(--radius-xs);
|
||||
color: var(--color-accent);
|
||||
}
|
||||
|
||||
.docs-article pre code {
|
||||
background: none;
|
||||
padding: 0;
|
||||
color: var(--color-code-text);
|
||||
}
|
||||
|
||||
/* ============================================
|
||||
Tables
|
||||
============================================ */
|
||||
|
||||
.docs-table {
|
||||
width: 100%;
|
||||
border-collapse: collapse;
|
||||
margin: var(--space-l) 0;
|
||||
font-size: var(--font-size-s);
|
||||
}
|
||||
|
||||
.docs-table th,
|
||||
.docs-table td {
|
||||
padding: var(--space-m);
|
||||
text-align: left;
|
||||
border-bottom: 1px solid var(--color-border);
|
||||
}
|
||||
|
||||
.docs-table th {
|
||||
font-weight: 600;
|
||||
color: var(--color-text-primary);
|
||||
background-color: var(--color-surface);
|
||||
}
|
||||
|
||||
.docs-table td {
|
||||
color: var(--color-text-secondary);
|
||||
}
|
||||
|
||||
.docs-table code {
|
||||
font-size: 0.85em;
|
||||
}
|
||||
|
||||
/* ============================================
|
||||
Callouts
|
||||
============================================ */
|
||||
|
||||
.callout {
|
||||
display: flex;
|
||||
gap: var(--space-m);
|
||||
padding: var(--space-l);
|
||||
border-radius: var(--radius-m);
|
||||
margin: var(--space-l) 0;
|
||||
}
|
||||
|
||||
.callout-icon {
|
||||
font-size: var(--font-size-l);
|
||||
flex-shrink: 0;
|
||||
}
|
||||
|
||||
.callout-content h4 {
|
||||
margin: 0 0 var(--space-s) 0;
|
||||
font-size: var(--font-size-s);
|
||||
}
|
||||
|
||||
.callout-content p {
|
||||
margin: 0;
|
||||
font-size: var(--font-size-s);
|
||||
}
|
||||
|
||||
.callout-info {
|
||||
background-color: rgba(99, 102, 241, 0.1);
|
||||
border: 1px solid rgba(99, 102, 241, 0.2);
|
||||
}
|
||||
|
||||
.callout-info .callout-icon {
|
||||
color: var(--color-accent);
|
||||
}
|
||||
|
||||
.callout-info .callout-content h4 {
|
||||
color: var(--color-accent);
|
||||
}
|
||||
|
||||
.callout-tip {
|
||||
background-color: rgba(16, 185, 129, 0.1);
|
||||
border: 1px solid rgba(16, 185, 129, 0.2);
|
||||
}
|
||||
|
||||
.callout-tip .callout-icon {
|
||||
color: var(--color-success);
|
||||
}
|
||||
|
||||
.callout-tip .callout-content h4 {
|
||||
color: var(--color-success);
|
||||
}
|
||||
|
||||
.callout-warning {
|
||||
background-color: rgba(245, 158, 11, 0.1);
|
||||
border: 1px solid rgba(245, 158, 11, 0.2);
|
||||
}
|
||||
|
||||
.callout-warning .callout-icon {
|
||||
color: var(--color-warning);
|
||||
}
|
||||
|
||||
.callout-warning .callout-content h4 {
|
||||
color: var(--color-warning);
|
||||
}
|
||||
|
||||
/* ============================================
|
||||
Badges
|
||||
============================================ */
|
||||
|
||||
.badge {
|
||||
display: inline-block;
|
||||
padding: 2px 8px;
|
||||
font-size: var(--font-size-xs);
|
||||
font-weight: 600;
|
||||
border-radius: var(--radius-s);
|
||||
text-transform: uppercase;
|
||||
letter-spacing: 0.03em;
|
||||
}
|
||||
|
||||
.badge-critical {
|
||||
background-color: rgba(239, 68, 68, 0.15);
|
||||
color: var(--color-error);
|
||||
}
|
||||
|
||||
.badge-important {
|
||||
background-color: rgba(245, 158, 11, 0.15);
|
||||
color: var(--color-warning);
|
||||
}
|
||||
|
||||
.badge-nice {
|
||||
background-color: rgba(99, 102, 241, 0.15);
|
||||
color: var(--color-accent);
|
||||
}
|
||||
|
||||
/* ============================================
|
||||
Philosophy Grid
|
||||
============================================ */
|
||||
|
||||
.philosophy-grid {
|
||||
display: grid;
|
||||
grid-template-columns: repeat(1, 1fr);
|
||||
gap: var(--space-l);
|
||||
margin: var(--space-xl) 0;
|
||||
}
|
||||
|
||||
@media (min-width: 640px) {
|
||||
.philosophy-grid {
|
||||
grid-template-columns: repeat(2, 1fr);
|
||||
}
|
||||
}
|
||||
|
||||
.philosophy-card {
|
||||
padding: var(--space-xl);
|
||||
background-color: var(--color-surface);
|
||||
border-radius: var(--radius-m);
|
||||
border: 1px solid var(--color-border);
|
||||
}
|
||||
|
||||
.philosophy-icon {
|
||||
font-size: var(--font-size-xl);
|
||||
color: var(--color-accent);
|
||||
margin-bottom: var(--space-m);
|
||||
}
|
||||
|
||||
.philosophy-card h4 {
|
||||
margin: 0 0 var(--space-s) 0;
|
||||
color: var(--color-text-primary);
|
||||
}
|
||||
|
||||
.philosophy-card p {
|
||||
margin: 0;
|
||||
font-size: var(--font-size-s);
|
||||
color: var(--color-text-secondary);
|
||||
}
|
||||
|
||||
/* ============================================
|
||||
Blockquotes
|
||||
============================================ */
|
||||
|
||||
.highlight-quote {
|
||||
font-size: var(--font-size-l);
|
||||
font-style: italic;
|
||||
color: var(--color-accent);
|
||||
padding: var(--space-xl);
|
||||
margin: var(--space-xl) 0;
|
||||
background: linear-gradient(135deg, var(--color-accent-lighter), transparent);
|
||||
border-left: 4px solid var(--color-accent);
|
||||
border-radius: var(--radius-m);
|
||||
}
|
||||
|
||||
/* ============================================
|
||||
Navigation Footer
|
||||
============================================ */
|
||||
|
||||
.docs-nav-footer {
|
||||
display: flex;
|
||||
justify-content: space-between;
|
||||
gap: var(--space-l);
|
||||
margin-top: var(--space-xxl);
|
||||
padding-top: var(--space-xl);
|
||||
border-top: 1px solid var(--color-border);
|
||||
}
|
||||
|
||||
.nav-prev,
|
||||
.nav-next {
|
||||
display: flex;
|
||||
flex-direction: column;
|
||||
gap: var(--space-xs);
|
||||
padding: var(--space-l);
|
||||
background-color: var(--color-surface);
|
||||
border-radius: var(--radius-m);
|
||||
text-decoration: none;
|
||||
transition: all 0.2s ease;
|
||||
flex: 1;
|
||||
max-width: 300px;
|
||||
}
|
||||
|
||||
.nav-prev:hover,
|
||||
.nav-next:hover {
|
||||
background-color: var(--color-surface-hover);
|
||||
border-color: var(--color-accent);
|
||||
}
|
||||
|
||||
.nav-next {
|
||||
text-align: right;
|
||||
margin-left: auto;
|
||||
}
|
||||
|
||||
.nav-label {
|
||||
font-size: var(--font-size-xs);
|
||||
color: var(--color-text-tertiary);
|
||||
text-transform: uppercase;
|
||||
letter-spacing: 0.05em;
|
||||
}
|
||||
|
||||
.nav-title {
|
||||
font-weight: 600;
|
||||
color: var(--color-accent);
|
||||
display: flex;
|
||||
align-items: center;
|
||||
gap: var(--space-s);
|
||||
}
|
||||
|
||||
.nav-next .nav-title {
|
||||
justify-content: flex-end;
|
||||
}
|
||||
|
||||
/* ============================================
|
||||
Mobile Sidebar Overlay
|
||||
============================================ */
|
||||
|
||||
@media (max-width: 1023px) {
|
||||
.docs-sidebar.open::before {
|
||||
content: '';
|
||||
position: fixed;
|
||||
top: 0;
|
||||
left: 0;
|
||||
right: 0;
|
||||
bottom: 0;
|
||||
background-color: rgba(0, 0, 0, 0.5);
|
||||
z-index: -1;
|
||||
}
|
||||
}
|
||||
|
||||
/* ============================================
|
||||
Changelog Styles
|
||||
============================================ */
|
||||
|
||||
.version-section {
|
||||
margin-bottom: var(--space-xxl);
|
||||
padding-bottom: var(--space-xl);
|
||||
border-bottom: 1px solid var(--color-border);
|
||||
}
|
||||
|
||||
.version-section:last-child {
|
||||
border-bottom: none;
|
||||
}
|
||||
|
||||
.version-header {
|
||||
display: flex;
|
||||
align-items: center;
|
||||
gap: var(--space-m);
|
||||
margin-bottom: var(--space-l);
|
||||
flex-wrap: wrap;
|
||||
}
|
||||
|
||||
.version-header h2 {
|
||||
margin: 0;
|
||||
padding: 0;
|
||||
border: none;
|
||||
font-size: var(--font-size-xl);
|
||||
color: var(--color-text-primary);
|
||||
}
|
||||
|
||||
.version-date {
|
||||
font-size: var(--font-size-s);
|
||||
color: var(--color-text-tertiary);
|
||||
background-color: var(--color-surface);
|
||||
padding: var(--space-xs) var(--space-m);
|
||||
border-radius: var(--radius-s);
|
||||
}
|
||||
|
||||
.version-badge {
|
||||
font-size: var(--font-size-xs);
|
||||
font-weight: 600;
|
||||
padding: var(--space-xs) var(--space-m);
|
||||
border-radius: var(--radius-s);
|
||||
background-color: var(--color-accent);
|
||||
color: white;
|
||||
}
|
||||
|
||||
.version-badge.major {
|
||||
background-color: var(--color-warning);
|
||||
}
|
||||
|
||||
.version-description {
|
||||
font-size: var(--font-size-m);
|
||||
color: var(--color-text-secondary);
|
||||
margin-bottom: var(--space-l);
|
||||
font-style: italic;
|
||||
}
|
||||
|
||||
.changelog-category {
|
||||
margin-bottom: var(--space-l);
|
||||
padding: var(--space-l);
|
||||
background-color: var(--color-surface);
|
||||
border-radius: var(--radius-m);
|
||||
border-left: 4px solid var(--color-border);
|
||||
}
|
||||
|
||||
.changelog-category h3 {
|
||||
margin: 0 0 var(--space-m) 0;
|
||||
font-size: var(--font-size-m);
|
||||
display: flex;
|
||||
align-items: center;
|
||||
gap: var(--space-s);
|
||||
}
|
||||
|
||||
.changelog-category h3 i {
|
||||
font-size: var(--font-size-s);
|
||||
}
|
||||
|
||||
.changelog-category h4 {
|
||||
margin: var(--space-l) 0 var(--space-s) 0;
|
||||
font-size: var(--font-size-s);
|
||||
color: var(--color-text-secondary);
|
||||
}
|
||||
|
||||
.changelog-category ul {
|
||||
margin: 0;
|
||||
padding-left: var(--space-xl);
|
||||
}
|
||||
|
||||
.changelog-category li {
|
||||
margin-bottom: var(--space-s);
|
||||
}
|
||||
|
||||
.changelog-category.added {
|
||||
border-left-color: var(--color-success);
|
||||
}
|
||||
|
||||
.changelog-category.added h3 {
|
||||
color: var(--color-success);
|
||||
}
|
||||
|
||||
.changelog-category.improved {
|
||||
border-left-color: var(--color-accent);
|
||||
}
|
||||
|
||||
.changelog-category.improved h3 {
|
||||
color: var(--color-accent);
|
||||
}
|
||||
|
||||
.changelog-category.changed {
|
||||
border-left-color: var(--color-warning);
|
||||
}
|
||||
|
||||
.changelog-category.changed h3 {
|
||||
color: var(--color-warning);
|
||||
}
|
||||
|
||||
.changelog-category.fixed {
|
||||
border-left-color: var(--color-error);
|
||||
}
|
||||
|
||||
.changelog-category.fixed h3 {
|
||||
color: var(--color-error);
|
||||
}
|
||||
|
||||
.version-summary {
|
||||
margin-top: var(--space-l);
|
||||
}
|
||||
|
||||
.version-summary h4 {
|
||||
margin-bottom: var(--space-m);
|
||||
}
|
||||
|
||||
.version-summary table {
|
||||
width: 100%;
|
||||
max-width: 400px;
|
||||
border-collapse: collapse;
|
||||
font-size: var(--font-size-s);
|
||||
}
|
||||
|
||||
.version-summary th,
|
||||
.version-summary td {
|
||||
padding: var(--space-s) var(--space-m);
|
||||
text-align: left;
|
||||
border-bottom: 1px solid var(--color-border);
|
||||
}
|
||||
|
||||
.version-summary th {
|
||||
font-weight: 600;
|
||||
background-color: var(--color-surface);
|
||||
}
|
||||
|
||||
.version-summary .positive {
|
||||
color: var(--color-success);
|
||||
font-weight: 600;
|
||||
}
|
||||
|
||||
.version-summary .negative {
|
||||
color: var(--color-error);
|
||||
font-weight: 600;
|
||||
}
|
||||
2886
docs/css/style.css
2886
docs/css/style.css
File diff suppressed because it is too large
Load Diff
1046
docs/index.html
1046
docs/index.html
File diff suppressed because it is too large
Load Diff
225
docs/js/main.js
225
docs/js/main.js
@@ -1,225 +0,0 @@
|
||||
/**
|
||||
* Compounding Engineering Documentation
|
||||
* Main JavaScript functionality
|
||||
*/
|
||||
|
||||
document.addEventListener('DOMContentLoaded', () => {
|
||||
initMobileNav();
|
||||
initSmoothScroll();
|
||||
initCopyCode();
|
||||
initThemeToggle();
|
||||
});
|
||||
|
||||
/**
|
||||
* Mobile Navigation Toggle
|
||||
*/
|
||||
function initMobileNav() {
|
||||
const mobileToggle = document.querySelector('[data-mobile-toggle]');
|
||||
const navigation = document.querySelector('[data-navigation]');
|
||||
|
||||
if (!mobileToggle || !navigation) return;
|
||||
|
||||
mobileToggle.addEventListener('click', () => {
|
||||
navigation.classList.toggle('open');
|
||||
mobileToggle.classList.toggle('active');
|
||||
|
||||
// Update aria-expanded
|
||||
const isOpen = navigation.classList.contains('open');
|
||||
mobileToggle.setAttribute('aria-expanded', isOpen);
|
||||
});
|
||||
|
||||
// Close menu when clicking outside
|
||||
document.addEventListener('click', (event) => {
|
||||
if (!mobileToggle.contains(event.target) && !navigation.contains(event.target)) {
|
||||
navigation.classList.remove('open');
|
||||
mobileToggle.classList.remove('active');
|
||||
mobileToggle.setAttribute('aria-expanded', 'false');
|
||||
}
|
||||
});
|
||||
|
||||
// Close menu when clicking a nav link
|
||||
navigation.querySelectorAll('.nav-link').forEach(link => {
|
||||
link.addEventListener('click', () => {
|
||||
navigation.classList.remove('open');
|
||||
mobileToggle.classList.remove('active');
|
||||
mobileToggle.setAttribute('aria-expanded', 'false');
|
||||
});
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* Smooth Scroll for Anchor Links
|
||||
*/
|
||||
function initSmoothScroll() {
|
||||
document.querySelectorAll('a[href^="#"]').forEach(anchor => {
|
||||
anchor.addEventListener('click', function(e) {
|
||||
const targetId = this.getAttribute('href');
|
||||
if (targetId === '#') return;
|
||||
|
||||
const targetElement = document.querySelector(targetId);
|
||||
if (!targetElement) return;
|
||||
|
||||
e.preventDefault();
|
||||
|
||||
const navHeight = document.querySelector('.nav-container')?.offsetHeight || 0;
|
||||
const targetPosition = targetElement.getBoundingClientRect().top + window.pageYOffset - navHeight - 24;
|
||||
|
||||
window.scrollTo({
|
||||
top: targetPosition,
|
||||
behavior: 'smooth'
|
||||
});
|
||||
|
||||
// Update URL without jumping
|
||||
history.pushState(null, null, targetId);
|
||||
});
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* Copy Code Functionality
|
||||
*/
|
||||
function initCopyCode() {
|
||||
document.querySelectorAll('.card-code-block').forEach(block => {
|
||||
// Create copy button
|
||||
const copyBtn = document.createElement('button');
|
||||
copyBtn.className = 'copy-btn';
|
||||
copyBtn.innerHTML = '<i class="fa-regular fa-copy"></i>';
|
||||
copyBtn.setAttribute('aria-label', 'Copy code');
|
||||
copyBtn.setAttribute('title', 'Copy to clipboard');
|
||||
|
||||
// Style the button
|
||||
copyBtn.style.cssText = `
|
||||
position: absolute;
|
||||
top: 8px;
|
||||
right: 8px;
|
||||
padding: 6px 10px;
|
||||
background: rgba(255, 255, 255, 0.1);
|
||||
border: none;
|
||||
border-radius: 6px;
|
||||
color: #94a3b8;
|
||||
cursor: pointer;
|
||||
opacity: 0;
|
||||
transition: all 0.2s ease;
|
||||
font-size: 14px;
|
||||
`;
|
||||
|
||||
// Make parent relative for positioning
|
||||
block.style.position = 'relative';
|
||||
block.appendChild(copyBtn);
|
||||
|
||||
// Show/hide on hover
|
||||
block.addEventListener('mouseenter', () => {
|
||||
copyBtn.style.opacity = '1';
|
||||
});
|
||||
|
||||
block.addEventListener('mouseleave', () => {
|
||||
copyBtn.style.opacity = '0';
|
||||
});
|
||||
|
||||
// Copy functionality
|
||||
copyBtn.addEventListener('click', async () => {
|
||||
const code = block.querySelector('code');
|
||||
if (!code) return;
|
||||
|
||||
try {
|
||||
await navigator.clipboard.writeText(code.textContent);
|
||||
copyBtn.innerHTML = '<i class="fa-solid fa-check"></i>';
|
||||
copyBtn.style.color = '#34d399';
|
||||
|
||||
setTimeout(() => {
|
||||
copyBtn.innerHTML = '<i class="fa-regular fa-copy"></i>';
|
||||
copyBtn.style.color = '#94a3b8';
|
||||
}, 2000);
|
||||
} catch (err) {
|
||||
console.error('Failed to copy:', err);
|
||||
copyBtn.innerHTML = '<i class="fa-solid fa-xmark"></i>';
|
||||
copyBtn.style.color = '#f87171';
|
||||
|
||||
setTimeout(() => {
|
||||
copyBtn.innerHTML = '<i class="fa-regular fa-copy"></i>';
|
||||
copyBtn.style.color = '#94a3b8';
|
||||
}, 2000);
|
||||
}
|
||||
});
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* Theme Toggle (Light/Dark)
|
||||
*/
|
||||
function initThemeToggle() {
|
||||
// Check for saved theme preference or default to dark
|
||||
const savedTheme = localStorage.getItem('theme') || 'dark';
|
||||
document.documentElement.className = `theme-${savedTheme}`;
|
||||
|
||||
// Create theme toggle button if it doesn't exist
|
||||
const existingToggle = document.querySelector('[data-theme-toggle]');
|
||||
if (existingToggle) {
|
||||
existingToggle.addEventListener('click', toggleTheme);
|
||||
updateThemeToggleIcon(existingToggle, savedTheme);
|
||||
}
|
||||
}
|
||||
|
||||
function toggleTheme() {
|
||||
const html = document.documentElement;
|
||||
const currentTheme = html.classList.contains('theme-dark') ? 'dark' : 'light';
|
||||
const newTheme = currentTheme === 'dark' ? 'light' : 'dark';
|
||||
|
||||
html.className = `theme-${newTheme}`;
|
||||
localStorage.setItem('theme', newTheme);
|
||||
|
||||
const toggle = document.querySelector('[data-theme-toggle]');
|
||||
if (toggle) {
|
||||
updateThemeToggleIcon(toggle, newTheme);
|
||||
}
|
||||
}
|
||||
|
||||
function updateThemeToggleIcon(toggle, theme) {
|
||||
const icon = toggle.querySelector('i');
|
||||
if (icon) {
|
||||
icon.className = theme === 'dark' ? 'fa-solid fa-sun' : 'fa-solid fa-moon';
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Intersection Observer for Animation on Scroll
|
||||
*/
|
||||
function initScrollAnimations() {
|
||||
const observerOptions = {
|
||||
threshold: 0.1,
|
||||
rootMargin: '0px 0px -50px 0px'
|
||||
};
|
||||
|
||||
const observer = new IntersectionObserver((entries) => {
|
||||
entries.forEach(entry => {
|
||||
if (entry.isIntersecting) {
|
||||
entry.target.classList.add('visible');
|
||||
observer.unobserve(entry.target);
|
||||
}
|
||||
});
|
||||
}, observerOptions);
|
||||
|
||||
document.querySelectorAll('.agent-card, .command-card, .skill-card, .mcp-card, .stat-card').forEach(card => {
|
||||
card.style.opacity = '0';
|
||||
card.style.transform = 'translateY(20px)';
|
||||
card.style.transition = 'opacity 0.5s ease, transform 0.5s ease';
|
||||
observer.observe(card);
|
||||
});
|
||||
}
|
||||
|
||||
// Add visible class styles
|
||||
const style = document.createElement('style');
|
||||
style.textContent = `
|
||||
.agent-card.visible,
|
||||
.command-card.visible,
|
||||
.skill-card.visible,
|
||||
.mcp-card.visible,
|
||||
.stat-card.visible {
|
||||
opacity: 1 !important;
|
||||
transform: translateY(0) !important;
|
||||
}
|
||||
`;
|
||||
document.head.appendChild(style);
|
||||
|
||||
// Initialize scroll animations after a short delay
|
||||
setTimeout(initScrollAnimations, 100);
|
||||
@@ -1,649 +0,0 @@
|
||||
<!DOCTYPE html>
|
||||
<html lang="en" class="theme-dark">
|
||||
|
||||
<head>
|
||||
<meta charset="utf-8" />
|
||||
<title>Agent Reference - Compounding Engineering</title>
|
||||
<meta content="Complete reference for all 23 specialized AI agents in the Compounding Engineering plugin." name="description" />
|
||||
<meta content="width=device-width, initial-scale=1" name="viewport" />
|
||||
|
||||
<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/font-awesome/6.5.1/css/all.min.css" />
|
||||
<link href="../css/style.css" rel="stylesheet" type="text/css" />
|
||||
<link href="../css/docs.css" rel="stylesheet" type="text/css" />
|
||||
<script src="../js/main.js" type="text/javascript" defer></script>
|
||||
</head>
|
||||
|
||||
<body>
|
||||
<div class="background-gradient"></div>
|
||||
<div class="docs-layout">
|
||||
<aside class="docs-sidebar">
|
||||
<div class="sidebar-header">
|
||||
<a href="../index.html" class="nav-brand">
|
||||
<span class="logo-icon"><i class="fa-solid fa-layer-group"></i></span>
|
||||
<span class="logo-text">CE Docs</span>
|
||||
</a>
|
||||
</div>
|
||||
<nav class="sidebar-nav">
|
||||
<div class="nav-section">
|
||||
<h3>Getting Started</h3>
|
||||
<ul>
|
||||
<li><a href="getting-started.html">Installation</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="nav-section">
|
||||
<h3>Reference</h3>
|
||||
<ul>
|
||||
<li><a href="agents.html" class="active">Agents (23)</a></li>
|
||||
<li><a href="commands.html">Commands (13)</a></li>
|
||||
<li><a href="skills.html">Skills (11)</a></li>
|
||||
<li><a href="mcp-servers.html">MCP Servers (2)</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="nav-section">
|
||||
<h3>Resources</h3>
|
||||
<ul>
|
||||
<li><a href="changelog.html">Changelog</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="nav-section">
|
||||
<h3>On This Page</h3>
|
||||
<ul>
|
||||
<li><a href="#review-agents">Review (10)</a></li>
|
||||
<li><a href="#research-agents">Research (4)</a></li>
|
||||
<li><a href="#workflow-agents">Workflow (5)</a></li>
|
||||
<li><a href="#design-agents">Design (3)</a></li>
|
||||
<li><a href="#docs-agents">Docs (1)</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
</nav>
|
||||
</aside>
|
||||
|
||||
<main class="docs-content">
|
||||
<div class="docs-header">
|
||||
<nav class="breadcrumb">
|
||||
<a href="../index.html">Home</a>
|
||||
<span>/</span>
|
||||
<a href="getting-started.html">Docs</a>
|
||||
<span>/</span>
|
||||
<span>Agents</span>
|
||||
</nav>
|
||||
<button class="mobile-menu-toggle" data-sidebar-toggle>
|
||||
<i class="fa-solid fa-bars"></i>
|
||||
</button>
|
||||
</div>
|
||||
|
||||
<article class="docs-article">
|
||||
<h1><i class="fa-solid fa-users-gear color-accent"></i> Agent Reference</h1>
|
||||
<p class="lead">
|
||||
Think of agents as your expert teammates who never sleep. You've got 23 specialists here—each one obsessed with a single domain. Call them individually when you need focused expertise, or orchestrate them together for multi-angle analysis. They're opinionated, they're fast, and they remember your codebase better than you do.
|
||||
</p>
|
||||
|
||||
<div class="usage-box">
|
||||
<h3>How to Use Agents</h3>
|
||||
<div class="card-code-block">
|
||||
<pre><code># Basic invocation
|
||||
claude agent [agent-name]
|
||||
|
||||
# With a specific message
|
||||
claude agent [agent-name] "Your message here"
|
||||
|
||||
# Examples
|
||||
claude agent kieran-rails-reviewer
|
||||
claude agent security-sentinel "Audit the payment flow"</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Review Agents -->
|
||||
<section id="review-agents">
|
||||
<h2><i class="fa-solid fa-code-pull-request"></i> Review Agents (10)</h2>
|
||||
<p>Your code review dream team. These agents catch what humans miss at 2am—security holes, performance cliffs, architectural drift, and those "it works but I hate it" moments. They're picky. They disagree with each other. That's the point.</p>
|
||||
|
||||
<div class="agent-detail" id="kieran-rails-reviewer">
|
||||
<div class="agent-detail-header">
|
||||
<h3>kieran-rails-reviewer</h3>
|
||||
<span class="agent-badge">Rails</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
Your senior Rails developer who's seen too many "clever" solutions fail in production. Obsessed with code that's boring, predictable, and maintainable. Strict on existing code (because touching it risks everything), pragmatic on new isolated features (because shipping matters). If you've ever thought "this works but feels wrong," this reviewer will tell you why.
|
||||
</p>
|
||||
<h4>Key Principles</h4>
|
||||
<ul>
|
||||
<li><strong>Existing Code Modifications</strong> - Very strict. Added complexity needs strong justification.</li>
|
||||
<li><strong>New Code</strong> - Pragmatic. If it's isolated and works, it's acceptable.</li>
|
||||
<li><strong>Turbo Streams</strong> - Simple turbo streams MUST be inline arrays in controllers.</li>
|
||||
<li><strong>Testing as Quality</strong> - Hard-to-test code = poor structure that needs refactoring.</li>
|
||||
<li><strong>Naming (5-Second Rule)</strong> - Must understand what a view/component does in 5 seconds from its name.</li>
|
||||
<li><strong>Namespacing</strong> - Always use <code>class Module::ClassName</code> pattern.</li>
|
||||
<li><strong>Duplication > Complexity</strong> - Simple duplicated code is better than complex DRY abstractions.</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent kieran-rails-reviewer "Review the UserController"</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="agent-detail" id="dhh-rails-reviewer">
|
||||
<div class="agent-detail-header">
|
||||
<h3>dhh-rails-reviewer</h3>
|
||||
<span class="agent-badge">Rails</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
What if DHH reviewed your Rails PR? He'd ask why you're building React inside Rails, why you need six layers of abstraction for a form, and whether you've forgotten that Rails already solved this problem. This agent channels that energy—blunt, opinionated, allergic to complexity.
|
||||
</p>
|
||||
<h4>Key Focus Areas</h4>
|
||||
<ul>
|
||||
<li>Identifies deviations from Rails conventions</li>
|
||||
<li>Spots JavaScript framework patterns infiltrating Rails</li>
|
||||
<li>Tears apart unnecessary abstractions</li>
|
||||
<li>Challenges overengineering and microservices mentality</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent dhh-rails-reviewer</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="agent-detail" id="kieran-python-reviewer">
|
||||
<div class="agent-detail-header">
|
||||
<h3>kieran-python-reviewer</h3>
|
||||
<span class="agent-badge">Python</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
Your Pythonic perfectionist who believes type hints aren't optional and <code>dict.get()</code> beats try/except KeyError. Expects modern Python 3.10+ patterns—no legacy syntax, no <code>typing.List</code> when <code>list</code> works natively. If your code looks like Java translated to Python, prepare for rewrites.
|
||||
</p>
|
||||
<h4>Key Focus Areas</h4>
|
||||
<ul>
|
||||
<li>Type hints for all functions</li>
|
||||
<li>Pythonic patterns and idioms</li>
|
||||
<li>Modern Python syntax</li>
|
||||
<li>Import organization</li>
|
||||
<li>Module extraction signals</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent kieran-python-reviewer</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="agent-detail" id="kieran-typescript-reviewer">
|
||||
<div class="agent-detail-header">
|
||||
<h3>kieran-typescript-reviewer</h3>
|
||||
<span class="agent-badge">TypeScript</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
TypeScript's type system is a gift—don't throw it away with <code>any</code>. This reviewer treats <code>any</code> like a code smell that needs justification. Expects proper types, clean imports, and code that doesn't need comments because the types explain everything. You added TypeScript for safety; this agent makes sure you actually get it.
|
||||
</p>
|
||||
<h4>Key Focus Areas</h4>
|
||||
<ul>
|
||||
<li>No <code>any</code> without justification</li>
|
||||
<li>Component/module extraction signals</li>
|
||||
<li>Import organization</li>
|
||||
<li>Modern TypeScript patterns</li>
|
||||
<li>Testability assessment</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent kieran-typescript-reviewer</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="agent-detail" id="security-sentinel">
|
||||
<div class="agent-detail-header">
|
||||
<h3>security-sentinel</h3>
|
||||
<span class="agent-badge critical">Security</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
Security vulnerabilities hide in boring code—the "just grab the user ID from params" line that ships a privilege escalation bug to production. This agent thinks like an attacker: SQL injection, XSS, auth bypass, leaked secrets. Run it before touching authentication, payments, or anything with PII. Your users' data depends on paranoia.
|
||||
</p>
|
||||
<h4>Security Checks</h4>
|
||||
<ul>
|
||||
<li>Input validation analysis</li>
|
||||
<li>SQL injection risk assessment</li>
|
||||
<li>XSS vulnerability detection</li>
|
||||
<li>Authentication/authorization audit</li>
|
||||
<li>Sensitive data exposure scanning</li>
|
||||
<li>OWASP Top 10 compliance</li>
|
||||
<li>Hardcoded secrets search</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent security-sentinel "Audit the payment flow"</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="agent-detail" id="performance-oracle">
|
||||
<div class="agent-detail-header">
|
||||
<h3>performance-oracle</h3>
|
||||
<span class="agent-badge">Performance</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
Your code works fine with 10 users. What happens at 10,000? This agent time-travels to your future scaling problems—N+1 queries that murder your database, O(n²) algorithms hiding in loops, missing indexes, memory leaks. It thinks in Big O notation and asks uncomfortable questions about what breaks first when traffic spikes.
|
||||
</p>
|
||||
<h4>Analysis Areas</h4>
|
||||
<ul>
|
||||
<li>Algorithmic complexity (Big O notation)</li>
|
||||
<li>N+1 query pattern detection</li>
|
||||
<li>Proper index usage verification</li>
|
||||
<li>Memory management review</li>
|
||||
<li>Caching opportunity identification</li>
|
||||
<li>Network usage optimization</li>
|
||||
<li>Frontend bundle impact</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent performance-oracle</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="agent-detail" id="architecture-strategist">
|
||||
<div class="agent-detail-header">
|
||||
<h3>architecture-strategist</h3>
|
||||
<span class="agent-badge">Architecture</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
Every "small change" either reinforces your architecture or starts eroding it. This agent zooms out to see if your fix actually fits the system's design—or if you're bolting duct tape onto a crumbling foundation. It speaks SOLID principles, microservice boundaries, and API contracts. Call it when you're about to make a change that "feels weird."
|
||||
</p>
|
||||
<h4>Analysis Areas</h4>
|
||||
<ul>
|
||||
<li>Overall system structure understanding</li>
|
||||
<li>Change context within architecture</li>
|
||||
<li>Architectural violation identification</li>
|
||||
<li>SOLID principles compliance</li>
|
||||
<li>Microservice boundary assessment</li>
|
||||
<li>API contract evaluation</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent architecture-strategist</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="agent-detail" id="data-integrity-guardian">
|
||||
<div class="agent-detail-header">
|
||||
<h3>data-integrity-guardian</h3>
|
||||
<span class="agent-badge critical">Data</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
Migrations can't be rolled back once they're run on production. This agent is your last line of defense before you accidentally drop a column with user data, create a race condition in transactions, or violate GDPR. It obsesses over referential integrity, rollback safety, and data constraints. Your database is forever; migrations should be paranoid.
|
||||
</p>
|
||||
<h4>Review Areas</h4>
|
||||
<ul>
|
||||
<li>Migration safety and reversibility</li>
|
||||
<li>Data constraint validation</li>
|
||||
<li>Transaction boundary review</li>
|
||||
<li>Referential integrity preservation</li>
|
||||
<li>Privacy compliance (GDPR, CCPA)</li>
|
||||
<li>Data corruption scenario checking</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent data-integrity-guardian</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="agent-detail" id="pattern-recognition-specialist">
|
||||
<div class="agent-detail-header">
|
||||
<h3>pattern-recognition-specialist</h3>
|
||||
<span class="agent-badge">Patterns</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
Patterns tell stories—Factory, Observer, God Object, Copy-Paste Programming. This agent reads your code like an archaeologist reading artifacts. It spots the good patterns (intentional design), the anti-patterns (accumulated tech debt), and the duplicated blocks you swore you'd refactor later. Runs tools like jscpd because humans miss repetition that machines catch instantly.
|
||||
</p>
|
||||
<h4>Detection Areas</h4>
|
||||
<ul>
|
||||
<li>Design patterns (Factory, Singleton, Observer, etc.)</li>
|
||||
<li>Anti-patterns and code smells</li>
|
||||
<li>TODO/FIXME comments</li>
|
||||
<li>God objects and circular dependencies</li>
|
||||
<li>Naming consistency</li>
|
||||
<li>Code duplication</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent pattern-recognition-specialist</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="agent-detail" id="code-simplicity-reviewer">
|
||||
<div class="agent-detail-header">
|
||||
<h3>code-simplicity-reviewer</h3>
|
||||
<span class="agent-badge">Quality</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
Simplicity is violent discipline. This agent asks "do you actually need this?" about every line, every abstraction, every dependency. YAGNI isn't a suggestion—it's the law. Your 200-line feature with three layers of indirection? This agent will show you the 50-line version that does the same thing. Complexity is a liability; simplicity compounds.
|
||||
</p>
|
||||
<h4>Simplification Checks</h4>
|
||||
<ul>
|
||||
<li>Analyze every line for necessity</li>
|
||||
<li>Simplify complex logic</li>
|
||||
<li>Remove redundancy and duplication</li>
|
||||
<li>Challenge abstractions</li>
|
||||
<li>Optimize for readability</li>
|
||||
<li>Eliminate premature generalization</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent code-simplicity-reviewer</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Research Agents -->
|
||||
<section id="research-agents">
|
||||
<h2><i class="fa-solid fa-microscope"></i> Research Agents (4)</h2>
|
||||
<p>Stop guessing. These agents dig through documentation, GitHub repos, git history, and real-world examples to give you answers backed by evidence. They read faster than you, remember more than you, and synthesize patterns you'd miss. Perfect for "how should I actually do this?" questions.</p>
|
||||
|
||||
<div class="agent-detail" id="framework-docs-researcher">
|
||||
<div class="agent-detail-header">
|
||||
<h3>framework-docs-researcher</h3>
|
||||
<span class="agent-badge">Research</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
Official docs are scattered. GitHub examples are inconsistent. Deprecations hide in changelogs. This agent pulls it all together—docs, source code, version constraints, real-world examples. Ask "how do I use Hotwire Turbo?" and get back patterns that actually work in production, not toy tutorials.
|
||||
</p>
|
||||
<h4>Capabilities</h4>
|
||||
<ul>
|
||||
<li>Fetch official framework and library documentation</li>
|
||||
<li>Identify version-specific constraints and deprecations</li>
|
||||
<li>Search GitHub for real-world usage examples</li>
|
||||
<li>Analyze gem/library source code using <code>bundle show</code></li>
|
||||
<li>Synthesize findings with practical examples</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent framework-docs-researcher "Research Hotwire Turbo patterns"</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="agent-detail" id="best-practices-researcher">
|
||||
<div class="agent-detail-header">
|
||||
<h3>best-practices-researcher</h3>
|
||||
<span class="agent-badge">Research</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
"Best practices" are everywhere and contradictory. This agent cuts through the noise by evaluating sources (official docs, trusted blogs, real GitHub repos), checking recency, and synthesizing actionable guidance. You get code templates, patterns that scale, and answers you can trust—not StackOverflow copy-paste roulette.
|
||||
</p>
|
||||
<h4>Capabilities</h4>
|
||||
<ul>
|
||||
<li>Leverage multiple sources (Context7 MCP, web search, GitHub)</li>
|
||||
<li>Evaluate information quality and recency</li>
|
||||
<li>Synthesize into actionable guidance</li>
|
||||
<li>Provide code examples and templates</li>
|
||||
<li>Research issue templates and community engagement</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent best-practices-researcher "Find pagination patterns"</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="agent-detail" id="git-history-analyzer">
|
||||
<div class="agent-detail-header">
|
||||
<h3>git-history-analyzer</h3>
|
||||
<span class="agent-badge">Git</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
Your codebase has a history—decisions, patterns, mistakes. This agent does archaeology with git tools: file evolution, blame analysis, contributor expertise mapping. Ask "why does this code exist?" and get the commit that explains it. Spot patterns in how bugs appear. Understand the design decisions buried in history.
|
||||
</p>
|
||||
<h4>Analysis Techniques</h4>
|
||||
<ul>
|
||||
<li>Trace file evolution using <code>git log --follow</code></li>
|
||||
<li>Determine code origins using <code>git blame -w -C -C -C</code></li>
|
||||
<li>Identify patterns from commit history</li>
|
||||
<li>Map key contributors and expertise areas</li>
|
||||
<li>Extract historical patterns of issues and fixes</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent git-history-analyzer "Analyze changes to User model"</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="agent-detail" id="repo-research-analyst">
|
||||
<div class="agent-detail-header">
|
||||
<h3>repo-research-analyst</h3>
|
||||
<span class="agent-badge">Research</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
Every repo has conventions—some documented, most tribal knowledge. This agent reads ARCHITECTURE.md, issue templates, PR patterns, and actual code to reverse-engineer the standards. Perfect for joining a new project or ensuring your PR matches the team's implicit style. Finds the rules nobody wrote down.
|
||||
</p>
|
||||
<h4>Analysis Areas</h4>
|
||||
<ul>
|
||||
<li>Architecture and documentation files (ARCHITECTURE.md, README.md, CLAUDE.md)</li>
|
||||
<li>GitHub issues for patterns and conventions</li>
|
||||
<li>Issue/PR templates and guidelines</li>
|
||||
<li>Implementation patterns using ast-grep or rg</li>
|
||||
<li>Project-specific conventions</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent repo-research-analyst</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Workflow Agents -->
|
||||
<section id="workflow-agents">
|
||||
<h2><i class="fa-solid fa-gears"></i> Workflow Agents (5)</h2>
|
||||
<p>Tedious work you hate doing. These agents handle the grind—reproducing bugs, resolving PR comments, running linters, analyzing specs. They're fast, they don't complain, and they free you up to solve interesting problems instead of mechanical ones.</p>
|
||||
|
||||
<div class="agent-detail" id="bug-reproduction-validator">
|
||||
<div class="agent-detail-header">
|
||||
<h3>bug-reproduction-validator</h3>
|
||||
<span class="agent-badge">Bugs</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
Half of bug reports aren't bugs—they're user errors, environment issues, or misunderstood features. This agent systematically reproduces the reported behavior, classifies what it finds (Confirmed, Can't Reproduce, Not a Bug, etc.), and assesses severity. Saves you from chasing ghosts or missing real issues.
|
||||
</p>
|
||||
<h4>Classification Types</h4>
|
||||
<ul>
|
||||
<li><strong>Confirmed</strong> - Bug reproduced successfully</li>
|
||||
<li><strong>Cannot Reproduce</strong> - Unable to reproduce</li>
|
||||
<li><strong>Not a Bug</strong> - Expected behavior</li>
|
||||
<li><strong>Environmental</strong> - Environment-specific issue</li>
|
||||
<li><strong>Data</strong> - Data-related issue</li>
|
||||
<li><strong>User Error</strong> - User misunderstanding</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent bug-reproduction-validator</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="agent-detail" id="pr-comment-resolver">
|
||||
<div class="agent-detail-header">
|
||||
<h3>pr-comment-resolver</h3>
|
||||
<span class="agent-badge">PR</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
Code review comments pile up. This agent reads them, plans fixes, implements changes, and reports back what it did. It doesn't argue with reviewers or skip hard feedback—it just resolves the work systematically. Great for burning through a dozen "change this variable name" comments in seconds.
|
||||
</p>
|
||||
<h4>Workflow</h4>
|
||||
<ul>
|
||||
<li>Analyze code review comments</li>
|
||||
<li>Plan the resolution before implementation</li>
|
||||
<li>Implement requested modifications</li>
|
||||
<li>Verify resolution doesn't break functionality</li>
|
||||
<li>Provide clear resolution reports</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent pr-comment-resolver</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="agent-detail" id="lint">
|
||||
<div class="agent-detail-header">
|
||||
<h3>lint</h3>
|
||||
<span class="agent-badge">Quality</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
Linters are pedantic robots that enforce consistency. This agent runs StandardRB, ERBLint, and Brakeman for you—checking Ruby style, ERB templates, and security issues. It's fast (uses the Haiku model) and catches the formatting noise before CI does.
|
||||
</p>
|
||||
<h4>Tools Run</h4>
|
||||
<ul>
|
||||
<li><code>bundle exec standardrb</code> - Ruby file checking/fixing</li>
|
||||
<li><code>bundle exec erblint --lint-all</code> - ERB templates</li>
|
||||
<li><code>bin/brakeman</code> - Security scanning</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent lint</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="agent-detail" id="spec-flow-analyzer">
|
||||
<div class="agent-detail-header">
|
||||
<h3>spec-flow-analyzer</h3>
|
||||
<span class="agent-badge">Testing</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
Specs always have gaps—edge cases nobody thought about, ambiguous requirements, missing error states. This agent maps all possible user flows, identifies what's unclear or missing, and generates the questions you need to ask stakeholders. Runs before you code to avoid building the wrong thing.
|
||||
</p>
|
||||
<h4>Analysis Areas</h4>
|
||||
<ul>
|
||||
<li>Map all possible user flows and permutations</li>
|
||||
<li>Identify gaps, ambiguities, and missing specifications</li>
|
||||
<li>Consider different user types, roles, permissions</li>
|
||||
<li>Analyze error states and edge cases</li>
|
||||
<li>Generate critical questions requiring clarification</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent spec-flow-analyzer</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="agent-detail" id="every-style-editor">
|
||||
<div class="agent-detail-header">
|
||||
<h3>every-style-editor</h3>
|
||||
<span class="agent-badge">Content</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
Style guides are arbitrary rules that make writing consistent. This agent enforces Every's particular quirks—title case in headlines, no overused filler words ("actually," "very"), active voice, Oxford commas. It's a line-by-line grammar cop for content that needs to match the brand.
|
||||
</p>
|
||||
<h4>Style Checks</h4>
|
||||
<ul>
|
||||
<li>Title case in headlines, sentence case elsewhere</li>
|
||||
<li>Company singular/plural usage</li>
|
||||
<li>Remove overused words (actually, very, just)</li>
|
||||
<li>Enforce active voice</li>
|
||||
<li>Apply formatting rules (Oxford commas, em dashes)</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent every-style-editor</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Design Agents -->
|
||||
<section id="design-agents">
|
||||
<h2><i class="fa-solid fa-palette"></i> Design Agents (3)</h2>
|
||||
<p>Design is iteration. These agents take screenshots, compare them to Figma, make targeted improvements, and repeat. They fix spacing, alignment, colors, typography—the visual details that compound into polish. Perfect for closing the gap between "it works" and "it looks right."</p>
|
||||
|
||||
<div class="agent-detail" id="design-iterator">
|
||||
<div class="agent-detail-header">
|
||||
<h3>design-iterator</h3>
|
||||
<span class="agent-badge">Design</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
Design doesn't happen in one pass. This agent runs a loop: screenshot the UI, analyze what's off (spacing, colors, alignment), implement 3-5 targeted fixes, repeat. Run it for 10 iterations and watch rough interfaces transform into polished designs through systematic refinement.
|
||||
</p>
|
||||
<h4>Process</h4>
|
||||
<ul>
|
||||
<li>Take focused screenshots of target elements</li>
|
||||
<li>Analyze current state and identify 3-5 improvements</li>
|
||||
<li>Implement targeted CSS/design changes</li>
|
||||
<li>Document changes made</li>
|
||||
<li>Repeat for specified iterations (default 10)</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent design-iterator</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="agent-detail" id="figma-design-sync">
|
||||
<div class="agent-detail-header">
|
||||
<h3>figma-design-sync</h3>
|
||||
<span class="agent-badge">Figma</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
Designers hand you a Figma file. You build it. Then: "the spacing is wrong, the font is off, the colors don't match." This agent compares your implementation to the Figma spec, identifies every visual discrepancy, and fixes them automatically. Designers stay happy. You stay sane.
|
||||
</p>
|
||||
<h4>Workflow</h4>
|
||||
<ul>
|
||||
<li>Extract design specifications from Figma</li>
|
||||
<li>Capture implementation screenshots</li>
|
||||
<li>Conduct systematic visual comparison</li>
|
||||
<li>Make precise code changes to fix discrepancies</li>
|
||||
<li>Verify implementation matches design</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent figma-design-sync</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="agent-detail" id="design-implementation-reviewer">
|
||||
<div class="agent-detail-header">
|
||||
<h3>design-implementation-reviewer</h3>
|
||||
<span class="agent-badge">Review</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
Before you ship UI changes, run this agent. It compares your implementation against Figma at a pixel level—layouts, typography, colors, spacing, responsive behavior. Uses the Opus model for detailed visual analysis. Catches the "close enough" mistakes that users notice but you don't.
|
||||
</p>
|
||||
<h4>Comparison Areas</h4>
|
||||
<ul>
|
||||
<li>Layouts and structure</li>
|
||||
<li>Typography (fonts, sizes, weights)</li>
|
||||
<li>Colors and themes</li>
|
||||
<li>Spacing and alignment</li>
|
||||
<li>Different viewport sizes</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent design-implementation-reviewer</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Docs Agents -->
|
||||
<section id="docs-agents">
|
||||
<h2><i class="fa-solid fa-file-lines"></i> Documentation Agent (1)</h2>
|
||||
|
||||
<div class="agent-detail" id="ankane-readme-writer">
|
||||
<div class="agent-detail-header">
|
||||
<h3>ankane-readme-writer</h3>
|
||||
<span class="agent-badge">Docs</span>
|
||||
</div>
|
||||
<p class="agent-detail-description">
|
||||
Andrew Kane writes READMEs that are models of clarity—concise, scannable, zero fluff. This agent generates gem documentation in that style: 15 words max per sentence, imperative voice, single-purpose code examples. If your README rambles, this agent will fix it.
|
||||
</p>
|
||||
<h4>Section Order</h4>
|
||||
<ol>
|
||||
<li>Header (title + description)</li>
|
||||
<li>Installation</li>
|
||||
<li>Quick Start</li>
|
||||
<li>Usage</li>
|
||||
<li>Options</li>
|
||||
<li>Upgrading</li>
|
||||
<li>Contributing</li>
|
||||
<li>License</li>
|
||||
</ol>
|
||||
<h4>Style Guidelines</h4>
|
||||
<ul>
|
||||
<li>Imperative voice throughout</li>
|
||||
<li>15 words max per sentence</li>
|
||||
<li>Single-purpose code fences</li>
|
||||
<li>Up to 4 badges maximum</li>
|
||||
<li>No HTML comments</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude agent ankane-readme-writer</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Navigation -->
|
||||
<nav class="docs-nav-footer">
|
||||
<a href="getting-started.html" class="nav-prev">
|
||||
<span class="nav-label">Previous</span>
|
||||
<span class="nav-title"><i class="fa-solid fa-arrow-left"></i> Getting Started</span>
|
||||
</a>
|
||||
<a href="commands.html" class="nav-next">
|
||||
<span class="nav-label">Next</span>
|
||||
<span class="nav-title">Commands <i class="fa-solid fa-arrow-right"></i></span>
|
||||
</a>
|
||||
</nav>
|
||||
</article>
|
||||
</main>
|
||||
</div>
|
||||
|
||||
<script>
|
||||
document.querySelector('[data-sidebar-toggle]')?.addEventListener('click', () => {
|
||||
document.querySelector('.docs-sidebar').classList.toggle('open');
|
||||
});
|
||||
</script>
|
||||
</body>
|
||||
|
||||
</html>
|
||||
@@ -1,534 +0,0 @@
|
||||
<!DOCTYPE html>
|
||||
<html lang="en" class="theme-dark">
|
||||
|
||||
<head>
|
||||
<meta charset="utf-8" />
|
||||
<title>Changelog - Compounding Engineering</title>
|
||||
<meta content="Version history and release notes for the Compounding Engineering plugin." name="description" />
|
||||
<meta content="width=device-width, initial-scale=1" name="viewport" />
|
||||
|
||||
<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/font-awesome/6.5.1/css/all.min.css" />
|
||||
<link href="../css/style.css" rel="stylesheet" type="text/css" />
|
||||
<link href="../css/docs.css" rel="stylesheet" type="text/css" />
|
||||
<script src="../js/main.js" type="text/javascript" defer></script>
|
||||
</head>
|
||||
|
||||
<body>
|
||||
<div class="background-gradient"></div>
|
||||
<div class="docs-layout">
|
||||
<aside class="docs-sidebar">
|
||||
<div class="sidebar-header">
|
||||
<a href="../index.html" class="nav-brand">
|
||||
<span class="logo-icon"><i class="fa-solid fa-layer-group"></i></span>
|
||||
<span class="logo-text">CE Docs</span>
|
||||
</a>
|
||||
</div>
|
||||
<nav class="sidebar-nav">
|
||||
<div class="nav-section">
|
||||
<h3>Getting Started</h3>
|
||||
<ul>
|
||||
<li><a href="getting-started.html">Installation</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="nav-section">
|
||||
<h3>Reference</h3>
|
||||
<ul>
|
||||
<li><a href="agents.html">Agents (23)</a></li>
|
||||
<li><a href="commands.html">Commands (13)</a></li>
|
||||
<li><a href="skills.html">Skills (11)</a></li>
|
||||
<li><a href="mcp-servers.html">MCP Servers (two)</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="nav-section">
|
||||
<h3>Resources</h3>
|
||||
<ul>
|
||||
<li><a href="changelog.html" class="active">Changelog</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="nav-section">
|
||||
<h3>On This Page</h3>
|
||||
<ul>
|
||||
<li><a href="#v2.32.2">v2.32.2</a></li>
|
||||
<li><a href="#v2.32.1">v2.32.1</a></li>
|
||||
<li><a href="#v2.6.0">v2.6.0</a></li>
|
||||
<li><a href="#v2.5.0">v2.5.0</a></li>
|
||||
<li><a href="#v2.4.1">v2.4.1</a></li>
|
||||
<li><a href="#v2.4.0">v2.4.0</a></li>
|
||||
<li><a href="#v2.3.0">v2.3.0</a></li>
|
||||
<li><a href="#v2.2.1">v2.2.1</a></li>
|
||||
<li><a href="#v2.2.0">v2.2.0</a></li>
|
||||
<li><a href="#v2.1.0">v2.1.0</a></li>
|
||||
<li><a href="#v2.0.0">v2.0.0</a></li>
|
||||
<li><a href="#v1.1.0">v1.1.0</a></li>
|
||||
<li><a href="#v1.0.0">v1.0.0</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
</nav>
|
||||
</aside>
|
||||
|
||||
<main class="docs-content">
|
||||
<div class="docs-header">
|
||||
<nav class="breadcrumb">
|
||||
<a href="../index.html">Home</a>
|
||||
<span>/</span>
|
||||
<a href="getting-started.html">Docs</a>
|
||||
<span>/</span>
|
||||
<span>Changelog</span>
|
||||
</nav>
|
||||
<button class="mobile-menu-toggle" data-sidebar-toggle>
|
||||
<i class="fa-solid fa-bars"></i>
|
||||
</button>
|
||||
</div>
|
||||
|
||||
<article class="docs-article">
|
||||
<h1><i class="fa-solid fa-clock-rotate-left color-accent"></i> Changelog</h1>
|
||||
<p class="lead">
|
||||
All notable changes to the compound-engineering plugin. This project follows
|
||||
<a href="https://semver.org/">Semantic Versioning</a> and
|
||||
<a href="https://keepachangelog.com/">Keep a Changelog</a> conventions.
|
||||
</p>
|
||||
|
||||
<!-- Version 2.32.2 -->
|
||||
<section id="v2.32.2" class="version-section">
|
||||
<div class="version-header">
|
||||
<h2>v2.32.2</h2>
|
||||
<span class="version-date">2026-02-12</span>
|
||||
</div>
|
||||
|
||||
<div class="changelog-category changed">
|
||||
<h3><i class="fa-solid fa-arrows-rotate"></i> Changed</h3>
|
||||
<ul>
|
||||
<li>
|
||||
<strong><code>/release-docs</code> command moved from plugin to local <code>.claude/commands/</code></strong> -
|
||||
This is a repository maintenance command and should not be distributed to users. Command count reduced from 24 to 23.
|
||||
</li>
|
||||
</ul>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Version 2.32.1 -->
|
||||
<section id="v2.32.1" class="version-section">
|
||||
<div class="version-header">
|
||||
<h2>v2.32.1</h2>
|
||||
<span class="version-date">2026-02-12</span>
|
||||
</div>
|
||||
|
||||
<div class="changelog-category changed">
|
||||
<h3><i class="fa-solid fa-arrows-rotate"></i> Changed</h3>
|
||||
<ul>
|
||||
<li>
|
||||
<strong><code>/workflows:review</code> command</strong> - Added <code>learnings-researcher</code>
|
||||
agent to the parallel review phase. The review now searches <code>docs/solutions/</code> for past
|
||||
issues related to the PR's modules and patterns, surfacing "Known Pattern" findings during synthesis.
|
||||
</li>
|
||||
</ul>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Version 2.6.0 -->
|
||||
<section id="v2.6.0" class="version-section">
|
||||
<div class="version-header">
|
||||
<h2>v2.6.0</h2>
|
||||
<span class="version-date">2024-11-26</span>
|
||||
</div>
|
||||
|
||||
<div class="changelog-category removed">
|
||||
<h3><i class="fa-solid fa-minus"></i> Removed</h3>
|
||||
<ul>
|
||||
<li>
|
||||
<strong><code>feedback-codifier</code> agent</strong> - Removed from workflow agents.
|
||||
Agent count reduced from 24 to 23.
|
||||
</li>
|
||||
</ul>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Version 2.5.0 -->
|
||||
<section id="v2.5.0" class="version-section">
|
||||
<div class="version-header">
|
||||
<h2>v2.5.0</h2>
|
||||
<span class="version-date">2024-11-25</span>
|
||||
</div>
|
||||
|
||||
<div class="changelog-category added">
|
||||
<h3><i class="fa-solid fa-plus"></i> Added</h3>
|
||||
<ul>
|
||||
<li>
|
||||
<strong><code>/report-bug</code> command</strong> - New slash command for reporting bugs in the
|
||||
compound-engineering plugin. Provides a structured workflow that gathers bug information
|
||||
through guided questions, collects environment details automatically, and creates a GitHub
|
||||
issue in the EveryInc/compound-engineering-plugin repository.
|
||||
</li>
|
||||
</ul>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Version 2.4.1 -->
|
||||
<section id="v2.4.1" class="version-section">
|
||||
<div class="version-header">
|
||||
<h2>v2.4.1</h2>
|
||||
<span class="version-date">2024-11-24</span>
|
||||
</div>
|
||||
|
||||
<div class="changelog-category improved">
|
||||
<h3><i class="fa-solid fa-arrow-up"></i> Improved</h3>
|
||||
<ul>
|
||||
<li>
|
||||
<strong>design-iterator agent</strong> - Added focused screenshot guidance: always capture
|
||||
only the target element/area instead of full page screenshots. Includes browser_resize
|
||||
recommendations, element-targeted screenshot workflow using browser_snapshot refs, and
|
||||
explicit instruction to never use fullPage mode.
|
||||
</li>
|
||||
</ul>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Version 2.4.0 -->
|
||||
<section id="v2.4.0" class="version-section">
|
||||
<div class="version-header">
|
||||
<h2>v2.4.0</h2>
|
||||
<span class="version-date">2024-11-24</span>
|
||||
</div>
|
||||
|
||||
<div class="changelog-category fixed">
|
||||
<h3><i class="fa-solid fa-bug"></i> Fixed</h3>
|
||||
<ul>
|
||||
<li>
|
||||
<strong>MCP Configuration</strong> - Moved MCP servers back to <code>plugin.json</code>
|
||||
following working examples from anthropics/life-sciences plugins.
|
||||
</li>
|
||||
<li>
|
||||
<strong>Context7 URL</strong> - Updated to use HTTP type with correct endpoint URL.
|
||||
</li>
|
||||
</ul>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Version 2.3.0 -->
|
||||
<section id="v2.3.0" class="version-section">
|
||||
<div class="version-header">
|
||||
<h2>v2.3.0</h2>
|
||||
<span class="version-date">2024-11-24</span>
|
||||
</div>
|
||||
|
||||
<div class="changelog-category changed">
|
||||
<h3><i class="fa-solid fa-arrows-rotate"></i> Changed</h3>
|
||||
<ul>
|
||||
<li>
|
||||
<strong>MCP Configuration</strong> - Moved MCP servers from inline <code>plugin.json</code>
|
||||
to separate <code>.mcp.json</code> file per Claude Code best practices.
|
||||
</li>
|
||||
</ul>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Version 2.2.1 -->
|
||||
<section id="v2.2.1" class="version-section">
|
||||
<div class="version-header">
|
||||
<h2>v2.2.1</h2>
|
||||
<span class="version-date">2024-11-24</span>
|
||||
</div>
|
||||
|
||||
<div class="changelog-category fixed">
|
||||
<h3><i class="fa-solid fa-bug"></i> Fixed</h3>
|
||||
<ul>
|
||||
<li>
|
||||
<strong>Playwright MCP Server</strong> - Added missing <code>"type": "stdio"</code> field
|
||||
required for MCP server configuration to load properly.
|
||||
</li>
|
||||
</ul>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Version 2.2.0 -->
|
||||
<section id="v2.2.0" class="version-section">
|
||||
<div class="version-header">
|
||||
<h2>v2.2.0</h2>
|
||||
<span class="version-date">2024-11-24</span>
|
||||
</div>
|
||||
|
||||
<div class="changelog-category added">
|
||||
<h3><i class="fa-solid fa-plus"></i> Added</h3>
|
||||
<ul>
|
||||
<li>
|
||||
<strong>Context7 MCP Server</strong> - Bundled Context7 for instant framework documentation
|
||||
lookup. Provides up-to-date docs for Rails, React, Next.js, and more than 100 other frameworks.
|
||||
</li>
|
||||
</ul>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Version 2.1.0 -->
|
||||
<section id="v2.1.0" class="version-section">
|
||||
<div class="version-header">
|
||||
<h2>v2.1.0</h2>
|
||||
<span class="version-date">2024-11-24</span>
|
||||
</div>
|
||||
|
||||
<div class="changelog-category added">
|
||||
<h3><i class="fa-solid fa-plus"></i> Added</h3>
|
||||
<ul>
|
||||
<li>
|
||||
<strong>Playwright MCP Server</strong> - Bundled <code>@playwright/mcp</code> for browser
|
||||
automation across all projects. Provides screenshot, navigation, click, fill, and evaluate tools.
|
||||
</li>
|
||||
</ul>
|
||||
</div>
|
||||
|
||||
<div class="changelog-category changed">
|
||||
<h3><i class="fa-solid fa-arrows-rotate"></i> Changed</h3>
|
||||
<ul>
|
||||
<li>Replaced all Puppeteer references with Playwright across agents and commands:
|
||||
<ul>
|
||||
<li><code>bug-reproduction-validator</code> agent</li>
|
||||
<li><code>design-iterator</code> agent</li>
|
||||
<li><code>design-implementation-reviewer</code> agent</li>
|
||||
<li><code>figma-design-sync</code> agent</li>
|
||||
<li><code>generate_command</code> command</li>
|
||||
</ul>
|
||||
</li>
|
||||
</ul>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Version 2.0.2 -->
|
||||
<section id="v2.0.2" class="version-section">
|
||||
<div class="version-header">
|
||||
<h2>v2.0.2</h2>
|
||||
<span class="version-date">2024-11-24</span>
|
||||
</div>
|
||||
|
||||
<div class="changelog-category changed">
|
||||
<h3><i class="fa-solid fa-arrows-rotate"></i> Changed</h3>
|
||||
<ul>
|
||||
<li>
|
||||
<strong>design-iterator agent</strong> - Updated description to emphasize proactive usage
|
||||
when design work isn't coming together on first attempt.
|
||||
</li>
|
||||
</ul>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Version 2.0.1 -->
|
||||
<section id="v2.0.1" class="version-section">
|
||||
<div class="version-header">
|
||||
<h2>v2.0.1</h2>
|
||||
<span class="version-date">2024-11-24</span>
|
||||
</div>
|
||||
|
||||
<div class="changelog-category added">
|
||||
<h3><i class="fa-solid fa-plus"></i> Added</h3>
|
||||
<ul>
|
||||
<li><strong>CLAUDE.md</strong> - Project instructions with versioning requirements</li>
|
||||
<li><strong>docs/solutions/plugin-versioning-requirements.md</strong> - Workflow documentation</li>
|
||||
</ul>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Version 2.0.0 -->
|
||||
<section id="v2.0.0" class="version-section">
|
||||
<div class="version-header">
|
||||
<h2>v2.0.0</h2>
|
||||
<span class="version-date">2024-11-24</span>
|
||||
<span class="version-badge major">Major Release</span>
|
||||
</div>
|
||||
|
||||
<p class="version-description">
|
||||
Major reorganization consolidating agents, commands, and skills from multiple sources into
|
||||
a single, well-organized plugin.
|
||||
</p>
|
||||
|
||||
<div class="changelog-category added">
|
||||
<h3><i class="fa-solid fa-plus"></i> Added</h3>
|
||||
|
||||
<h4>New Agents (seven)</h4>
|
||||
<ul>
|
||||
<li><code>design-iterator</code> - Iteratively refine UI components through systematic design iterations</li>
|
||||
<li><code>design-implementation-reviewer</code> - Verify UI implementations match Figma design specifications</li>
|
||||
<li><code>figma-design-sync</code> - Synchronize web implementations with Figma designs</li>
|
||||
<li><code>bug-reproduction-validator</code> - Systematically reproduce and validate bug reports</li>
|
||||
<li><code>spec-flow-analyzer</code> - Analyze user flows and identify gaps in specifications</li>
|
||||
<li><code>lint</code> - Run linting and code quality checks on Ruby and ERB files</li>
|
||||
<li><code>ankane-readme-writer</code> - Create READMEs following Ankane-style template for Ruby gems</li>
|
||||
</ul>
|
||||
|
||||
<h4>New Commands (nine)</h4>
|
||||
<ul>
|
||||
<li><code>/changelog</code> - Create engaging changelogs for recent merges</li>
|
||||
<li><code>/plan_review</code> - Multi-agent plan review in parallel</li>
|
||||
<li><code>/resolve_parallel</code> - Resolve TODO comments in parallel</li>
|
||||
<li><code>/resolve_pr_parallel</code> - Resolve PR comments in parallel</li>
|
||||
<li><code>/reproduce-bug</code> - Reproduce bugs using logs and console</li>
|
||||
<li><code>/prime</code> - Prime/setup command</li>
|
||||
<li><code>/create-agent-skill</code> - Create or edit Claude Code skills</li>
|
||||
<li><code>/heal-skill</code> - Fix skill documentation issues</li>
|
||||
<li><code>/codify</code> - Document solved problems for knowledge base</li>
|
||||
</ul>
|
||||
|
||||
<h4>New Skills (10)</h4>
|
||||
<ul>
|
||||
<li><code>andrew-kane-gem-writer</code> - Write Ruby gems following Andrew Kane's patterns</li>
|
||||
<li><code>codify-docs</code> - Capture solved problems as categorized documentation</li>
|
||||
<li><code>create-agent-skills</code> - Expert guidance for creating Claude Code skills</li>
|
||||
<li><code>dhh-ruby-style</code> - Write Ruby/Rails code in DHH's 37signals style</li>
|
||||
<li><code>dspy-ruby</code> - Build type-safe LLM applications with DSPy.rb</li>
|
||||
<li><code>every-style-editor</code> - Review copy for Every's style guide compliance</li>
|
||||
<li><code>file-todos</code> - File-based todo tracking system</li>
|
||||
<li><code>frontend-design</code> - Create production-grade frontend interfaces</li>
|
||||
<li><code>git-worktree</code> - Manage Git worktrees for parallel development</li>
|
||||
<li><code>skill-creator</code> - Guide for creating effective Claude Code skills</li>
|
||||
</ul>
|
||||
</div>
|
||||
|
||||
<div class="changelog-category changed">
|
||||
<h3><i class="fa-solid fa-arrows-rotate"></i> Changed</h3>
|
||||
<h4>Agents Reorganized by Category</h4>
|
||||
<ul>
|
||||
<li><code>review/</code> (10 agents) - Code quality, security, performance reviewers</li>
|
||||
<li><code>research/</code> (four agents) - Documentation, patterns, history analysis</li>
|
||||
<li><code>design/</code> (three agents) - UI/design review and iteration</li>
|
||||
<li><code>workflow/</code> (six agents) - PR resolution, bug validation, linting</li>
|
||||
<li><code>docs/</code> (one agent) - README generation</li>
|
||||
</ul>
|
||||
</div>
|
||||
|
||||
<div class="version-summary">
|
||||
<h4>Summary</h4>
|
||||
<table>
|
||||
<thead>
|
||||
<tr>
|
||||
<th>Component</th>
|
||||
<th>v1.1.0</th>
|
||||
<th>v2.0.0</th>
|
||||
<th>Change</th>
|
||||
</tr>
|
||||
</thead>
|
||||
<tbody>
|
||||
<tr>
|
||||
<td>Agents</td>
|
||||
<td>17</td>
|
||||
<td>24</td>
|
||||
<td class="positive">+7</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td>Commands</td>
|
||||
<td>6</td>
|
||||
<td>15</td>
|
||||
<td class="positive">+9</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td>Skills</td>
|
||||
<td>1</td>
|
||||
<td>11</td>
|
||||
<td class="positive">+10</td>
|
||||
</tr>
|
||||
</tbody>
|
||||
</table>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Version 1.1.0 -->
|
||||
<section id="v1.1.0" class="version-section">
|
||||
<div class="version-header">
|
||||
<h2>v1.1.0</h2>
|
||||
<span class="version-date">2024-11-22</span>
|
||||
</div>
|
||||
|
||||
<div class="changelog-category added">
|
||||
<h3><i class="fa-solid fa-plus"></i> Added</h3>
|
||||
<ul>
|
||||
<li>
|
||||
<strong>gemini-imagegen Skill</strong>
|
||||
<ul>
|
||||
<li>Text-to-image generation with Google's Gemini API</li>
|
||||
<li>Image editing and manipulation</li>
|
||||
<li>Multi-turn refinement via chat interface</li>
|
||||
<li>Multiple reference image composition (up to 14 images)</li>
|
||||
<li>Model support: <code>gemini-2.5-flash-image</code> and <code>gemini-3-pro-image-preview</code></li>
|
||||
</ul>
|
||||
</li>
|
||||
</ul>
|
||||
</div>
|
||||
|
||||
<div class="changelog-category fixed">
|
||||
<h3><i class="fa-solid fa-bug"></i> Fixed</h3>
|
||||
<ul>
|
||||
<li>Corrected component counts in documentation (17 agents, not 15)</li>
|
||||
</ul>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Version 1.0.0 -->
|
||||
<section id="v1.0.0" class="version-section">
|
||||
<div class="version-header">
|
||||
<h2>v1.0.0</h2>
|
||||
<span class="version-date">2024-10-09</span>
|
||||
<span class="version-badge">Initial Release</span>
|
||||
</div>
|
||||
|
||||
<p class="version-description">
|
||||
Initial release of the compound-engineering plugin.
|
||||
</p>
|
||||
|
||||
<div class="changelog-category added">
|
||||
<h3><i class="fa-solid fa-plus"></i> Added</h3>
|
||||
|
||||
<h4>17 Specialized Agents</h4>
|
||||
|
||||
<p><strong>Code Review (five)</strong></p>
|
||||
<ul>
|
||||
<li><code>kieran-rails-reviewer</code> - Rails code review with strict conventions</li>
|
||||
<li><code>kieran-python-reviewer</code> - Python code review with quality standards</li>
|
||||
<li><code>kieran-typescript-reviewer</code> - TypeScript code review</li>
|
||||
<li><code>dhh-rails-reviewer</code> - Rails review from DHH's perspective</li>
|
||||
<li><code>code-simplicity-reviewer</code> - Final pass for simplicity and minimalism</li>
|
||||
</ul>
|
||||
|
||||
<p><strong>Analysis & Architecture (four)</strong></p>
|
||||
<ul>
|
||||
<li><code>architecture-strategist</code> - Architectural decisions and compliance</li>
|
||||
<li><code>pattern-recognition-specialist</code> - Design pattern analysis</li>
|
||||
<li><code>security-sentinel</code> - Security audits and vulnerability assessments</li>
|
||||
<li><code>performance-oracle</code> - Performance analysis and optimization</li>
|
||||
</ul>
|
||||
|
||||
<p><strong>Research (four)</strong></p>
|
||||
<ul>
|
||||
<li><code>framework-docs-researcher</code> - Framework documentation research</li>
|
||||
<li><code>best-practices-researcher</code> - External best practices gathering</li>
|
||||
<li><code>git-history-analyzer</code> - Git history and code evolution analysis</li>
|
||||
<li><code>repo-research-analyst</code> - Repository structure and conventions</li>
|
||||
</ul>
|
||||
|
||||
<p><strong>Workflow (three)</strong></p>
|
||||
<ul>
|
||||
<li><code>every-style-editor</code> - Every's style guide compliance</li>
|
||||
<li><code>pr-comment-resolver</code> - PR comment resolution</li>
|
||||
<li><code>feedback-codifier</code> - Feedback pattern codification</li>
|
||||
</ul>
|
||||
|
||||
<h4>Six Slash Commands</h4>
|
||||
<ul>
|
||||
<li><code>/plan</code> - Create implementation plans</li>
|
||||
<li><code>/review</code> - Comprehensive code reviews</li>
|
||||
<li><code>/work</code> - Execute work items systematically</li>
|
||||
<li><code>/triage</code> - Triage and prioritize issues</li>
|
||||
<li><code>/resolve_todo_parallel</code> - Resolve TODOs in parallel</li>
|
||||
<li><code>/generate_command</code> - Generate new slash commands</li>
|
||||
</ul>
|
||||
|
||||
<h4>Infrastructure</h4>
|
||||
<ul>
|
||||
<li>MIT license</li>
|
||||
<li>Plugin manifest (<code>plugin.json</code>)</li>
|
||||
<li>Pre-configured permissions for Rails development</li>
|
||||
</ul>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
</article>
|
||||
</main>
|
||||
</div>
|
||||
</body>
|
||||
|
||||
</html>
|
||||
@@ -1,523 +0,0 @@
|
||||
<!DOCTYPE html>
|
||||
<html lang="en" class="theme-dark">
|
||||
|
||||
<head>
|
||||
<meta charset="utf-8" />
|
||||
<title>Command Reference - Compounding Engineering</title>
|
||||
<meta content="Complete reference for all 16 slash commands in the Compounding Engineering plugin." name="description" />
|
||||
<meta content="width=device-width, initial-scale=1" name="viewport" />
|
||||
|
||||
<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/font-awesome/6.5.1/css/all.min.css" />
|
||||
<link href="../css/style.css" rel="stylesheet" type="text/css" />
|
||||
<link href="../css/docs.css" rel="stylesheet" type="text/css" />
|
||||
<script src="../js/main.js" type="text/javascript" defer></script>
|
||||
</head>
|
||||
|
||||
<body>
|
||||
<div class="background-gradient"></div>
|
||||
<div class="docs-layout">
|
||||
<aside class="docs-sidebar">
|
||||
<div class="sidebar-header">
|
||||
<a href="../index.html" class="nav-brand">
|
||||
<span class="logo-icon"><i class="fa-solid fa-layer-group"></i></span>
|
||||
<span class="logo-text">CE Docs</span>
|
||||
</a>
|
||||
</div>
|
||||
<nav class="sidebar-nav">
|
||||
<div class="nav-section">
|
||||
<h3>Getting Started</h3>
|
||||
<ul>
|
||||
<li><a href="getting-started.html">Installation</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="nav-section">
|
||||
<h3>Reference</h3>
|
||||
<ul>
|
||||
<li><a href="agents.html">Agents (23)</a></li>
|
||||
<li><a href="commands.html" class="active">Commands (13)</a></li>
|
||||
<li><a href="skills.html">Skills (11)</a></li>
|
||||
<li><a href="mcp-servers.html">MCP Servers (two)</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="nav-section">
|
||||
<h3>Resources</h3>
|
||||
<ul>
|
||||
<li><a href="changelog.html">Changelog</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="nav-section">
|
||||
<h3>On This Page</h3>
|
||||
<ul>
|
||||
<li><a href="#workflow-commands">Workflow (four)</a></li>
|
||||
<li><a href="#utility-commands">Utility (12)</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
</nav>
|
||||
</aside>
|
||||
|
||||
<main class="docs-content">
|
||||
<div class="docs-header">
|
||||
<nav class="breadcrumb">
|
||||
<a href="../index.html">Home</a>
|
||||
<span>/</span>
|
||||
<a href="getting-started.html">Docs</a>
|
||||
<span>/</span>
|
||||
<span>Commands</span>
|
||||
</nav>
|
||||
<button class="mobile-menu-toggle" data-sidebar-toggle>
|
||||
<i class="fa-solid fa-bars"></i>
|
||||
</button>
|
||||
</div>
|
||||
|
||||
<article class="docs-article">
|
||||
<h1><i class="fa-solid fa-terminal color-accent"></i> Command Reference</h1>
|
||||
<p class="lead">
|
||||
Here's the thing about slash commands: they're workflows you'd spend 20 minutes doing manually, compressed into one line. Type <code>/plan</code> and watch three agents launch in parallel to research your codebase while you grab coffee. That's the point—automation that actually saves time, not busywork dressed up as productivity.
|
||||
</p>
|
||||
|
||||
<!-- Workflow Commands -->
|
||||
<section id="workflow-commands">
|
||||
<h2><i class="fa-solid fa-arrows-spin"></i> Workflow Commands (four)</h2>
|
||||
<p>These are the big four: Plan your feature, Review your code, Work through the implementation, and Codify what you learned. Every professional developer does this cycle—these commands just make you faster at it.</p>
|
||||
|
||||
<div class="command-detail" id="workflows-plan">
|
||||
<div class="command-detail-header">
|
||||
<code class="command-detail-name">/plan</code>
|
||||
</div>
|
||||
<p class="command-detail-description">
|
||||
You've got a feature request and a blank page. This command turns "we need OAuth" into a structured plan that actually tells you what to build—researched, reviewed, and ready to execute.
|
||||
</p>
|
||||
<h4>Arguments</h4>
|
||||
<p><code>[feature description, bug report, or improvement idea]</code></p>
|
||||
<h4>Workflow</h4>
|
||||
<ol>
|
||||
<li><strong>Repository Research (Parallel)</strong> - Launch three agents simultaneously:
|
||||
<ul>
|
||||
<li><code>repo-research-analyst</code> - Project patterns</li>
|
||||
<li><code>best-practices-researcher</code> - Industry standards</li>
|
||||
<li><code>framework-docs-researcher</code> - Framework documentation</li>
|
||||
</ul>
|
||||
</li>
|
||||
<li><strong>SpecFlow Analysis</strong> - Run <code>spec-flow-analyzer</code> for user flows</li>
|
||||
<li><strong>Choose Detail Level</strong>:
|
||||
<ul>
|
||||
<li><strong>MINIMAL</strong> - Simple bugs/small improvements</li>
|
||||
<li><strong>MORE</strong> - Standard features</li>
|
||||
<li><strong>A LOT</strong> - Major features with phases</li>
|
||||
</ul>
|
||||
</li>
|
||||
<li><strong>Write Plan</strong> - Save as <code>plans/<issue_title>.md</code></li>
|
||||
<li><strong>Review</strong> - Call <code>/plan_review</code> for multi-agent feedback</li>
|
||||
</ol>
|
||||
<div class="callout callout-info">
|
||||
<div class="callout-icon"><i class="fa-solid fa-circle-info"></i></div>
|
||||
<div class="callout-content">
|
||||
<p>This command does NOT write code. It only researches and creates the plan.</p>
|
||||
</div>
|
||||
</div>
|
||||
<div class="card-code-block">
|
||||
<pre><code>/plan Add OAuth integration for third-party auth
|
||||
/plan Fix N+1 query in user dashboard</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="command-detail" id="workflows-review">
|
||||
<div class="command-detail-header">
|
||||
<code class="command-detail-name">/review</code>
|
||||
</div>
|
||||
<p class="command-detail-description">
|
||||
Twelve specialized reviewers examine your PR in parallel—security, performance, architecture, patterns. It's like code review by committee, except the committee finishes in two minutes instead of two days.
|
||||
</p>
|
||||
<h4>Arguments</h4>
|
||||
<p><code>[PR number, GitHub URL, branch name, or "latest"]</code></p>
|
||||
<h4>Workflow</h4>
|
||||
<ol>
|
||||
<li><strong>Setup</strong> - Detect review target, optionally use git-worktree for isolation</li>
|
||||
<li><strong>Launch 12 Parallel Review Agents</strong>:
|
||||
<ul>
|
||||
<li><code>kieran-rails-reviewer</code>, <code>dhh-rails-reviewer</code></li>
|
||||
<li><code>security-sentinel</code>, <code>performance-oracle</code></li>
|
||||
<li><code>architecture-strategist</code>, <code>data-integrity-guardian</code></li>
|
||||
<li><code>pattern-recognition-specialist</code>, <code>git-history-analyzer</code></li>
|
||||
<li>And more...</li>
|
||||
</ul>
|
||||
</li>
|
||||
<li><strong>Ultra-Thinking Analysis</strong> - Stakeholder perspectives, scenario exploration</li>
|
||||
<li><strong>Simplification Review</strong> - Run <code>code-simplicity-reviewer</code></li>
|
||||
<li><strong>Synthesize Findings</strong> - Categorize by severity (P1/P2/P3)</li>
|
||||
<li><strong>Create Todo Files</strong> - Using file-todos skill for all findings</li>
|
||||
</ol>
|
||||
<div class="callout callout-warning">
|
||||
<div class="callout-icon"><i class="fa-solid fa-triangle-exclamation"></i></div>
|
||||
<div class="callout-content">
|
||||
<p><strong>P1 (Critical) findings BLOCK MERGE.</strong> Address these before merging.</p>
|
||||
</div>
|
||||
</div>
|
||||
<div class="card-code-block">
|
||||
<pre><code>/review 42
|
||||
/review https://github.com/owner/repo/pull/42
|
||||
/review feature-branch-name
|
||||
/review latest</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="command-detail" id="workflows-work">
|
||||
<div class="command-detail-header">
|
||||
<code class="command-detail-name">/work</code>
|
||||
</div>
|
||||
<p class="command-detail-description">
|
||||
Point this at a plan file and watch it execute—reading requirements, setting up environment, running tests, creating commits, opening PRs. It's the "just build the thing" button you wish you always had.
|
||||
</p>
|
||||
<h4>Arguments</h4>
|
||||
<p><code>[plan file, specification, or todo file path]</code></p>
|
||||
<h4>Phases</h4>
|
||||
<ol>
|
||||
<li><strong>Quick Start</strong>
|
||||
<ul>
|
||||
<li>Read plan & clarify requirements</li>
|
||||
<li>Setup environment (live or worktree)</li>
|
||||
<li>Create TodoWrite task list</li>
|
||||
</ul>
|
||||
</li>
|
||||
<li><strong>Execute</strong>
|
||||
<ul>
|
||||
<li>Task execution loop with progress tracking</li>
|
||||
<li>Follow existing patterns</li>
|
||||
<li>Test continuously</li>
|
||||
<li>Figma sync if applicable</li>
|
||||
</ul>
|
||||
</li>
|
||||
<li><strong>Quality Check</strong>
|
||||
<ul>
|
||||
<li>Run test suite</li>
|
||||
<li>Run linting</li>
|
||||
<li>Optional reviewer agents for complex changes</li>
|
||||
</ul>
|
||||
</li>
|
||||
<li><strong>Ship It</strong>
|
||||
<ul>
|
||||
<li>Create commit with conventional format</li>
|
||||
<li>Create pull request</li>
|
||||
<li>Notify with summary</li>
|
||||
</ul>
|
||||
</li>
|
||||
</ol>
|
||||
<div class="card-code-block">
|
||||
<pre><code>/work plans/user-authentication.md
|
||||
/work todos/042-ready-p1-performance-issue.md</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="command-detail" id="workflows-compound">
|
||||
<div class="command-detail-header">
|
||||
<code class="command-detail-name">/compound</code>
|
||||
</div>
|
||||
<p class="command-detail-description">
|
||||
Just fixed a gnarly bug? This captures the solution before you forget it. Seven agents analyze what you did, why it worked, and how to prevent it next time. Each documented solution compounds your team's knowledge.
|
||||
</p>
|
||||
<h4>Arguments</h4>
|
||||
<p><code>[optional: brief context about the fix]</code></p>
|
||||
<h4>Workflow</h4>
|
||||
<ol>
|
||||
<li><strong>Preconditions</strong> - Verify problem is solved and verified working</li>
|
||||
<li><strong>Launch seven parallel subagents</strong>:
|
||||
<ul>
|
||||
<li>Context Analyzer - Extract YAML frontmatter skeleton</li>
|
||||
<li>Solution Extractor - Identify root cause and solution</li>
|
||||
<li>Related Docs Finder - Find cross-references</li>
|
||||
<li>Prevention Strategist - Develop prevention strategies</li>
|
||||
<li>Category Classifier - Determine docs category</li>
|
||||
<li>Documentation Writer - Create the file</li>
|
||||
<li>Optional Specialized Agent - Based on problem type</li>
|
||||
</ul>
|
||||
</li>
|
||||
<li><strong>Create Documentation</strong> - File in <code>docs/solutions/[category]/</code></li>
|
||||
</ol>
|
||||
<h4>Auto-Triggers</h4>
|
||||
<p>Phrases: "that worked", "it's fixed", "working now", "problem solved"</p>
|
||||
<div class="card-code-block">
|
||||
<pre><code>/compound
|
||||
/compound N+1 query optimization</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Utility Commands -->
|
||||
<section id="utility-commands">
|
||||
<h2><i class="fa-solid fa-wrench"></i> Utility Commands (12)</h2>
|
||||
<p>The supporting cast—commands that do one specific thing really well. Generate changelogs, resolve todos in parallel, triage findings, create new commands. The utilities you reach for daily.</p>
|
||||
|
||||
<div class="command-detail" id="changelog">
|
||||
<div class="command-detail-header">
|
||||
<code class="command-detail-name">/changelog</code>
|
||||
</div>
|
||||
<p class="command-detail-description">
|
||||
Turn your git history into a changelog people actually want to read. Breaking changes at the top, fun facts at the bottom, everything organized by what matters to your users.
|
||||
</p>
|
||||
<h4>Arguments</h4>
|
||||
<p><code>[optional: daily|weekly, or time period in days]</code></p>
|
||||
<h4>Output Sections</h4>
|
||||
<ul>
|
||||
<li>Breaking Changes (top priority)</li>
|
||||
<li>New Features</li>
|
||||
<li>Bug Fixes</li>
|
||||
<li>Other Improvements</li>
|
||||
<li>Shoutouts</li>
|
||||
<li>Fun Fact</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>/changelog daily
|
||||
/changelog weekly
|
||||
/changelog 7</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="command-detail" id="create-agent-skill">
|
||||
<div class="command-detail-header">
|
||||
<code class="command-detail-name">/create-agent-skill</code>
|
||||
</div>
|
||||
<p class="command-detail-description">
|
||||
Need a new skill? This walks you through creating one that actually works—proper frontmatter, clear documentation, all the conventions baked in. Think of it as scaffolding for skills.
|
||||
</p>
|
||||
<h4>Arguments</h4>
|
||||
<p><code>[skill description or requirements]</code></p>
|
||||
<div class="card-code-block">
|
||||
<pre><code>/create-agent-skill PDF processing for document analysis
|
||||
/create-agent-skill Web scraping with error handling</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="command-detail" id="generate-command">
|
||||
<div class="command-detail-header">
|
||||
<code class="command-detail-name">/generate_command</code>
|
||||
</div>
|
||||
<p class="command-detail-description">
|
||||
Same idea, but for commands instead of skills. Tell it what workflow you're tired of doing manually, and it generates a proper slash command with all the right patterns.
|
||||
</p>
|
||||
<h4>Arguments</h4>
|
||||
<p><code>[command purpose and requirements]</code></p>
|
||||
<div class="card-code-block">
|
||||
<pre><code>/generate_command Security audit for codebase
|
||||
/generate_command Automated performance testing</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="command-detail" id="heal-skill">
|
||||
<div class="command-detail-header">
|
||||
<code class="command-detail-name">/heal-skill</code>
|
||||
</div>
|
||||
<p class="command-detail-description">
|
||||
Skills drift—APIs change, URLs break, parameters get renamed. When a skill stops working, this figures out what's wrong and fixes the documentation. You approve the changes before anything commits.
|
||||
</p>
|
||||
<h4>Arguments</h4>
|
||||
<p><code>[optional: specific issue to fix]</code></p>
|
||||
<h4>Approval Options</h4>
|
||||
<ol>
|
||||
<li>Apply and commit</li>
|
||||
<li>Apply without commit</li>
|
||||
<li>Revise changes</li>
|
||||
<li>Cancel</li>
|
||||
</ol>
|
||||
<div class="card-code-block">
|
||||
<pre><code>/heal-skill API endpoint URL changed
|
||||
/heal-skill parameter validation error</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="command-detail" id="plan-review">
|
||||
<div class="command-detail-header">
|
||||
<code class="command-detail-name">/plan_review</code>
|
||||
</div>
|
||||
<p class="command-detail-description">
|
||||
Before you execute a plan, have three reviewers tear it apart—Rails conventions, best practices, simplicity. Better to find the problems in the plan than in production.
|
||||
</p>
|
||||
<h4>Arguments</h4>
|
||||
<p><code>[plan file path or plan content]</code></p>
|
||||
<h4>Review Agents</h4>
|
||||
<ul>
|
||||
<li><code>dhh-rails-reviewer</code> - Rails conventions</li>
|
||||
<li><code>kieran-rails-reviewer</code> - Rails best practices</li>
|
||||
<li><code>code-simplicity-reviewer</code> - Simplicity and clarity</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>/plan_review plans/user-authentication.md</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="command-detail" id="report-bug">
|
||||
<div class="command-detail-header">
|
||||
<code class="command-detail-name">/report-bug</code>
|
||||
</div>
|
||||
<p class="command-detail-description">
|
||||
Something broken? This collects all the context—what broke, what you expected, error messages, environment—and files a proper bug report. No more "it doesn't work" issues.
|
||||
</p>
|
||||
<h4>Arguments</h4>
|
||||
<p><code>[optional: brief description of the bug]</code></p>
|
||||
<h4>Information Collected</h4>
|
||||
<ul>
|
||||
<li>Bug category (Agent/Command/Skill/MCP/Installation)</li>
|
||||
<li>Specific component name</li>
|
||||
<li>Actual vs expected behavior</li>
|
||||
<li>Steps to reproduce</li>
|
||||
<li>Error messages</li>
|
||||
<li>Environment info (auto-gathered)</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>/report-bug Agent not working
|
||||
/report-bug Command failing with timeout</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="command-detail" id="reproduce-bug">
|
||||
<div class="command-detail-header">
|
||||
<code class="command-detail-name">/reproduce-bug</code>
|
||||
</div>
|
||||
<p class="command-detail-description">
|
||||
Give it a GitHub issue number and it tries to actually reproduce the bug—reading the issue, analyzing code paths, iterating until it finds the root cause. Then it posts findings back to the issue.
|
||||
</p>
|
||||
<h4>Arguments</h4>
|
||||
<p><code>[GitHub issue number]</code></p>
|
||||
<h4>Investigation Process</h4>
|
||||
<ol>
|
||||
<li>Read GitHub issue details</li>
|
||||
<li>Launch parallel investigation agents</li>
|
||||
<li>Analyze code for failure points</li>
|
||||
<li>Iterate until root cause found</li>
|
||||
<li>Post findings to GitHub issue</li>
|
||||
</ol>
|
||||
<div class="card-code-block">
|
||||
<pre><code>/reproduce-bug 142</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="command-detail" id="triage">
|
||||
<div class="command-detail-header">
|
||||
<code class="command-detail-name">/triage</code>
|
||||
</div>
|
||||
<p class="command-detail-description">
|
||||
Got a pile of code review findings or security audit results? This turns them into actionable todos—one at a time, you decide: create the todo, skip it, or modify and re-present.
|
||||
</p>
|
||||
<h4>Arguments</h4>
|
||||
<p><code>[findings list or source type]</code></p>
|
||||
<h4>User Decisions</h4>
|
||||
<ul>
|
||||
<li><strong>"yes"</strong> - Create/update todo file, change status to ready</li>
|
||||
<li><strong>"next"</strong> - Skip and delete from todos</li>
|
||||
<li><strong>"custom"</strong> - Modify and re-present</li>
|
||||
</ul>
|
||||
<div class="callout callout-info">
|
||||
<div class="callout-icon"><i class="fa-solid fa-circle-info"></i></div>
|
||||
<div class="callout-content">
|
||||
<p>This command does NOT write code. It only categorizes and creates todo files.</p>
|
||||
</div>
|
||||
</div>
|
||||
<div class="card-code-block">
|
||||
<pre><code>/triage code-review-findings.txt
|
||||
/triage security-audit-results</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="command-detail" id="resolve-parallel">
|
||||
<div class="command-detail-header">
|
||||
<code class="command-detail-name">/resolve_parallel</code>
|
||||
</div>
|
||||
<p class="command-detail-description">
|
||||
All those TODO comments scattered through your codebase? This finds them, builds a dependency graph, and spawns parallel agents to resolve them all at once. Clears the backlog in minutes.
|
||||
</p>
|
||||
<h4>Arguments</h4>
|
||||
<p><code>[optional: specific TODO pattern or file]</code></p>
|
||||
<h4>Process</h4>
|
||||
<ol>
|
||||
<li>Analyze TODO comments from codebase</li>
|
||||
<li>Create dependency graph (mermaid diagram)</li>
|
||||
<li>Spawn parallel <code>pr-comment-resolver</code> agents</li>
|
||||
<li>Commit and push after completion</li>
|
||||
</ol>
|
||||
<div class="card-code-block">
|
||||
<pre><code>/resolve_parallel
|
||||
/resolve_parallel authentication
|
||||
/resolve_parallel src/auth/</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="command-detail" id="resolve-pr-parallel">
|
||||
<div class="command-detail-header">
|
||||
<code class="command-detail-name">/resolve_pr_parallel</code>
|
||||
</div>
|
||||
<p class="command-detail-description">
|
||||
Same deal, but for PR review comments. Fetch unresolved threads, spawn parallel resolver agents, commit the fixes, and mark threads as resolved. Your reviewers will wonder how you're so fast.
|
||||
</p>
|
||||
<h4>Arguments</h4>
|
||||
<p><code>[optional: PR number or current PR]</code></p>
|
||||
<h4>Process</h4>
|
||||
<ol>
|
||||
<li>Get all unresolved PR comments</li>
|
||||
<li>Create TodoWrite list</li>
|
||||
<li>Launch parallel <code>pr-comment-resolver</code> agents</li>
|
||||
<li>Commit, resolve threads, and push</li>
|
||||
</ol>
|
||||
<div class="card-code-block">
|
||||
<pre><code>/resolve_pr_parallel
|
||||
/resolve_pr_parallel 123</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="command-detail" id="resolve-todo-parallel">
|
||||
<div class="command-detail-header">
|
||||
<code class="command-detail-name">/resolve_todo_parallel</code>
|
||||
</div>
|
||||
<p class="command-detail-description">
|
||||
Those todo files in your <code>/todos</code> directory? Point this at them and watch parallel agents knock them out—analyzing dependencies, executing in the right order, marking resolved as they finish.
|
||||
</p>
|
||||
<h4>Arguments</h4>
|
||||
<p><code>[optional: specific todo ID or pattern]</code></p>
|
||||
<h4>Process</h4>
|
||||
<ol>
|
||||
<li>Get unresolved TODOs from <code>/todos/*.md</code></li>
|
||||
<li>Analyze dependencies</li>
|
||||
<li>Spawn parallel agents</li>
|
||||
<li>Commit, mark as resolved, push</li>
|
||||
</ol>
|
||||
<div class="card-code-block">
|
||||
<pre><code>/resolve_todo_parallel
|
||||
/resolve_todo_parallel 042
|
||||
/resolve_todo_parallel p1</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="command-detail" id="prime">
|
||||
<div class="command-detail-header">
|
||||
<code class="command-detail-name">/prime</code>
|
||||
</div>
|
||||
<p class="command-detail-description">
|
||||
Your project initialization command. What exactly it does depends on your project setup—think of it as the "get everything ready" button before you start coding.
|
||||
</p>
|
||||
<div class="card-code-block">
|
||||
<pre><code>/prime</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Navigation -->
|
||||
<nav class="docs-nav-footer">
|
||||
<a href="agents.html" class="nav-prev">
|
||||
<span class="nav-label">Previous</span>
|
||||
<span class="nav-title"><i class="fa-solid fa-arrow-left"></i> Agents</span>
|
||||
</a>
|
||||
<a href="skills.html" class="nav-next">
|
||||
<span class="nav-label">Next</span>
|
||||
<span class="nav-title">Skills <i class="fa-solid fa-arrow-right"></i></span>
|
||||
</a>
|
||||
</nav>
|
||||
</article>
|
||||
</main>
|
||||
</div>
|
||||
|
||||
<script>
|
||||
document.querySelector('[data-sidebar-toggle]')?.addEventListener('click', () => {
|
||||
document.querySelector('.docs-sidebar').classList.toggle('open');
|
||||
});
|
||||
</script>
|
||||
</body>
|
||||
|
||||
</html>
|
||||
@@ -1,582 +0,0 @@
|
||||
<!DOCTYPE html>
|
||||
<html lang="en" class="theme-dark">
|
||||
|
||||
<head>
|
||||
<meta charset="utf-8" />
|
||||
<title>Getting Started - Compounding Engineering</title>
|
||||
<meta content="Complete guide to installing and using the Compounding Engineering plugin for Claude Code." name="description" />
|
||||
<meta content="width=device-width, initial-scale=1" name="viewport" />
|
||||
|
||||
<!-- Styles -->
|
||||
<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/font-awesome/6.5.1/css/all.min.css" />
|
||||
<link href="../css/style.css" rel="stylesheet" type="text/css" />
|
||||
<link href="../css/docs.css" rel="stylesheet" type="text/css" />
|
||||
<script src="../js/main.js" type="text/javascript" defer></script>
|
||||
</head>
|
||||
|
||||
<body>
|
||||
<div class="background-gradient"></div>
|
||||
<div class="docs-layout">
|
||||
<!-- Sidebar -->
|
||||
<aside class="docs-sidebar">
|
||||
<div class="sidebar-header">
|
||||
<a href="../index.html" class="nav-brand">
|
||||
<span class="logo-icon"><i class="fa-solid fa-layer-group"></i></span>
|
||||
<span class="logo-text">CE Docs</span>
|
||||
</a>
|
||||
</div>
|
||||
<nav class="sidebar-nav">
|
||||
<div class="nav-section">
|
||||
<h3>Getting Started</h3>
|
||||
<ul>
|
||||
<li><a href="#installation" class="active">Installation</a></li>
|
||||
<li><a href="#quick-start">Quick Start</a></li>
|
||||
<li><a href="#configuration">Configuration</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="nav-section">
|
||||
<h3>Core Concepts</h3>
|
||||
<ul>
|
||||
<li><a href="#philosophy">Philosophy</a></li>
|
||||
<li><a href="#agents">Using Agents</a></li>
|
||||
<li><a href="#commands">Using Commands</a></li>
|
||||
<li><a href="#skills">Using Skills</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="nav-section">
|
||||
<h3>Guides</h3>
|
||||
<ul>
|
||||
<li><a href="#code-review">Code Review Workflow</a></li>
|
||||
<li><a href="#creating-agents">Creating Custom Agents</a></li>
|
||||
<li><a href="#creating-skills">Creating Custom Skills</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="nav-section">
|
||||
<h3>Reference</h3>
|
||||
<ul>
|
||||
<li><a href="agents.html">Agent Reference</a></li>
|
||||
<li><a href="commands.html">Command Reference</a></li>
|
||||
<li><a href="skills.html">Skill Reference</a></li>
|
||||
<li><a href="mcp-servers.html">MCP Servers</a></li>
|
||||
<li><a href="changelog.html">Changelog</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
</nav>
|
||||
</aside>
|
||||
|
||||
<!-- Main Content -->
|
||||
<main class="docs-content">
|
||||
<div class="docs-header">
|
||||
<nav class="breadcrumb">
|
||||
<a href="../index.html">Home</a>
|
||||
<span>/</span>
|
||||
<span>Getting Started</span>
|
||||
</nav>
|
||||
<button class="mobile-menu-toggle" data-sidebar-toggle>
|
||||
<i class="fa-solid fa-bars"></i>
|
||||
</button>
|
||||
</div>
|
||||
|
||||
<article class="docs-article">
|
||||
<h1>Getting Started with Compounding Engineering</h1>
|
||||
<p class="lead">
|
||||
Five minutes from now, you'll run a single command that spins up 10 AI agents—each with a different specialty—to review your pull request in parallel. Security, performance, architecture, accessibility, all happening at once. That's the plugin. Let's get you set up.
|
||||
</p>
|
||||
|
||||
<!-- Installation Section -->
|
||||
<section id="installation">
|
||||
<h2><i class="fa-solid fa-download"></i> Installation</h2>
|
||||
|
||||
<h3>Prerequisites</h3>
|
||||
<ul>
|
||||
<li><a href="https://claude.ai/claude-code" target="_blank">Claude Code</a> installed and configured</li>
|
||||
<li>A GitHub account (for marketplace access)</li>
|
||||
<li>Node.js 18+ (for MCP servers)</li>
|
||||
</ul>
|
||||
|
||||
<h3>Step 1: Add the Marketplace</h3>
|
||||
<p>Think of the marketplace as an app store. You're adding it to Claude Code's list of places to look for plugins:</p>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude /plugin marketplace add https://github.com/EveryInc/compound-engineering-plugin</code></pre>
|
||||
</div>
|
||||
|
||||
<h3>Step 2: Install the Plugin</h3>
|
||||
<p>Now grab the plugin itself:</p>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude /plugin install compound-engineering</code></pre>
|
||||
</div>
|
||||
|
||||
<h3>Step 3: Verify Installation</h3>
|
||||
<p>Check that it worked:</p>
|
||||
<div class="card-code-block">
|
||||
<pre><code>claude /plugin list</code></pre>
|
||||
</div>
|
||||
<p>You'll see <code>compound-engineering</code> in the list. If you do, you're ready.</p>
|
||||
|
||||
<div class="callout callout-info">
|
||||
<div class="callout-icon"><i class="fa-solid fa-circle-info"></i></div>
|
||||
<div class="callout-content">
|
||||
<h4>Known Issue: MCP Servers</h4>
|
||||
<p>
|
||||
The bundled MCP servers (Playwright for browser automation, Context7 for docs) don't always auto-load. If you need them, there's a manual config step below. Otherwise, ignore this—everything else works fine.
|
||||
</p>
|
||||
</div>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Quick Start Section -->
|
||||
<section id="quick-start">
|
||||
<h2><i class="fa-solid fa-rocket"></i> Quick Start</h2>
|
||||
|
||||
<p>Let's see what this thing can actually do. I'll show you three workflows you'll use constantly:</p>
|
||||
|
||||
<h3>Run a Code Review</h3>
|
||||
<p>This is the big one. Type <code>/review</code> and watch it spawn 10+ specialized reviewers:</p>
|
||||
<div class="card-code-block">
|
||||
<pre><code># Review a PR by number
|
||||
/review 123
|
||||
|
||||
# Review the current branch
|
||||
/review
|
||||
|
||||
# Review a specific branch
|
||||
/review feature/my-feature</code></pre>
|
||||
</div>
|
||||
|
||||
<h3>Use a Specialized Agent</h3>
|
||||
<p>Sometimes you just need one expert. Call them directly:</p>
|
||||
<div class="card-code-block">
|
||||
<pre><code># Rails code review with Kieran's conventions
|
||||
claude agent kieran-rails-reviewer "Review the UserController"
|
||||
|
||||
# Security audit
|
||||
claude agent security-sentinel "Audit authentication flow"
|
||||
|
||||
# Research best practices
|
||||
claude agent best-practices-researcher "Find pagination patterns for Rails"</code></pre>
|
||||
</div>
|
||||
|
||||
<h3>Invoke a Skill</h3>
|
||||
<p>Skills are like loading a reference book into Claude's brain. When you need deep knowledge in a specific domain:</p>
|
||||
<div class="card-code-block">
|
||||
<pre><code># Generate images with Gemini
|
||||
skill: gemini-imagegen
|
||||
|
||||
# Write Ruby in DHH's style
|
||||
skill: dhh-ruby-style
|
||||
|
||||
# Create a new Claude Code skill
|
||||
skill: create-agent-skills</code></pre>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Configuration Section -->
|
||||
<section id="configuration">
|
||||
<h2><i class="fa-solid fa-gear"></i> Configuration</h2>
|
||||
|
||||
<h3 id="mcp-configuration">MCP Server Configuration</h3>
|
||||
<p>
|
||||
If the MCP servers didn't load automatically, paste this into <code>.claude/settings.json</code>:
|
||||
</p>
|
||||
<div class="card-code-block">
|
||||
<pre><code>{
|
||||
"mcpServers": {
|
||||
"playwright": {
|
||||
"type": "stdio",
|
||||
"command": "npx",
|
||||
"args": ["-y", "@playwright/mcp@latest"],
|
||||
"env": {}
|
||||
},
|
||||
"context7": {
|
||||
"type": "http",
|
||||
"url": "https://mcp.context7.com/mcp"
|
||||
}
|
||||
}
|
||||
}</code></pre>
|
||||
</div>
|
||||
|
||||
<h3>Environment Variables</h3>
|
||||
<p>Right now, only one skill needs an API key. If you use Gemini's image generation:</p>
|
||||
<table class="docs-table">
|
||||
<thead>
|
||||
<tr>
|
||||
<th>Variable</th>
|
||||
<th>Required For</th>
|
||||
<th>Description</th>
|
||||
</tr>
|
||||
</thead>
|
||||
<tbody>
|
||||
<tr>
|
||||
<td><code>GEMINI_API_KEY</code></td>
|
||||
<td>gemini-imagegen</td>
|
||||
<td>Google Gemini API key for image generation</td>
|
||||
</tr>
|
||||
</tbody>
|
||||
</table>
|
||||
</section>
|
||||
|
||||
<!-- Philosophy Section -->
|
||||
<section id="philosophy">
|
||||
<h2><i class="fa-solid fa-lightbulb"></i> The Compounding Engineering Philosophy</h2>
|
||||
|
||||
<blockquote class="highlight-quote">
|
||||
Every unit of engineering work should make subsequent units of work easier—not harder.
|
||||
</blockquote>
|
||||
|
||||
<p>Here's how it works in practice—the four-step loop you'll run over and over:</p>
|
||||
|
||||
<div class="philosophy-grid">
|
||||
<div class="philosophy-card">
|
||||
<div class="philosophy-icon"><i class="fa-solid fa-brain"></i></div>
|
||||
<h4>1. Plan</h4>
|
||||
<p>
|
||||
Before you write a single line, figure out what you're building and why. Use research agents to gather examples, patterns, and context. Think of it as Google Search meets expert consultation.
|
||||
</p>
|
||||
</div>
|
||||
<div class="philosophy-card">
|
||||
<div class="philosophy-icon"><i class="fa-solid fa-robot"></i></div>
|
||||
<h4>2. Delegate</h4>
|
||||
<p>
|
||||
Now build it—with help. Each agent specializes in something (Rails, security, design). You stay in the driver's seat, but you've got a team of specialists riding shotgun.
|
||||
</p>
|
||||
</div>
|
||||
<div class="philosophy-card">
|
||||
<div class="philosophy-icon"><i class="fa-solid fa-magnifying-glass"></i></div>
|
||||
<h4>3. Assess</h4>
|
||||
<p>
|
||||
Before you ship, run the gauntlet. Security agent checks for vulnerabilities. Performance agent flags N+1 queries. Architecture agent questions your design choices. All at once, all in parallel.
|
||||
</p>
|
||||
</div>
|
||||
<div class="philosophy-card">
|
||||
<div class="philosophy-icon"><i class="fa-solid fa-book"></i></div>
|
||||
<h4>4. Codify</h4>
|
||||
<p>
|
||||
You just solved a problem. Write it down. Next time you (or your teammate) face this, you'll have a runbook. That's the "compounding" part—each solution makes the next one faster.
|
||||
</p>
|
||||
</div>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Using Agents Section -->
|
||||
<section id="agents">
|
||||
<h2><i class="fa-solid fa-users-gear"></i> Using Agents</h2>
|
||||
|
||||
<p>
|
||||
Think of agents as coworkers with different job titles. You wouldn't ask your security engineer to design your UI, right? Same concept here—each agent has a specialty, and you call the one you need.
|
||||
</p>
|
||||
|
||||
<h3>Invoking Agents</h3>
|
||||
<div class="card-code-block">
|
||||
<pre><code># Basic syntax
|
||||
claude agent [agent-name] "[optional message]"
|
||||
|
||||
# Examples
|
||||
claude agent kieran-rails-reviewer
|
||||
claude agent security-sentinel "Audit the payment flow"
|
||||
claude agent git-history-analyzer "Show changes to user model"</code></pre>
|
||||
</div>
|
||||
|
||||
<h3>Agent Categories</h3>
|
||||
<table class="docs-table">
|
||||
<thead>
|
||||
<tr>
|
||||
<th>Category</th>
|
||||
<th>Count</th>
|
||||
<th>Purpose</th>
|
||||
</tr>
|
||||
</thead>
|
||||
<tbody>
|
||||
<tr>
|
||||
<td>Review</td>
|
||||
<td>10</td>
|
||||
<td>Code review, security audits, performance analysis</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td>Research</td>
|
||||
<td>four</td>
|
||||
<td>Best practices, documentation, git history</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td>Design</td>
|
||||
<td>three</td>
|
||||
<td>UI iteration, Figma sync, design review</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td>Workflow</td>
|
||||
<td>five</td>
|
||||
<td>Bug reproduction, PR resolution, linting</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td>Docs</td>
|
||||
<td>one</td>
|
||||
<td>README generation</td>
|
||||
</tr>
|
||||
</tbody>
|
||||
</table>
|
||||
|
||||
<p>
|
||||
<a href="agents.html" class="button secondary">
|
||||
<i class="fa-solid fa-arrow-right"></i> View All Agents
|
||||
</a>
|
||||
</p>
|
||||
</section>
|
||||
|
||||
<!-- Using Commands Section -->
|
||||
<section id="commands">
|
||||
<h2><i class="fa-solid fa-terminal"></i> Using Commands</h2>
|
||||
|
||||
<p>
|
||||
Commands are macros that run entire workflows for you. One command can spin up a dozen agents, coordinate their work, collect results, and hand you a summary. It's automation all the way down.
|
||||
</p>
|
||||
|
||||
<h3>Running Commands</h3>
|
||||
<div class="card-code-block">
|
||||
<pre><code># Workflow commands
|
||||
/plan
|
||||
/review 123
|
||||
/work
|
||||
/compound
|
||||
|
||||
# Utility commands
|
||||
/changelog
|
||||
/triage
|
||||
/reproduce-bug</code></pre>
|
||||
</div>
|
||||
|
||||
<h3>The Review Workflow</h3>
|
||||
<p>Let me show you what happens when you run <code>/review</code>. Here's the sequence:</p>
|
||||
<ol>
|
||||
<li><strong>Detection</strong> - Figures out what you want reviewed (PR number, branch name, or current changes)</li>
|
||||
<li><strong>Isolation</strong> - Spins up a git worktree so the review doesn't mess with your working directory</li>
|
||||
<li><strong>Parallel execution</strong> - Launches 10+ agents simultaneously (security, performance, architecture, accessibility...)</li>
|
||||
<li><strong>Synthesis</strong> - Sorts findings by severity (P1 = blocks merge, P2 = should fix, P3 = nice-to-have)</li>
|
||||
<li><strong>Persistence</strong> - Creates todo files so you don't lose track of issues</li>
|
||||
<li><strong>Summary</strong> - Hands you a readable report with action items</li>
|
||||
</ol>
|
||||
|
||||
<p>
|
||||
<a href="commands.html" class="button secondary">
|
||||
<i class="fa-solid fa-arrow-right"></i> View All Commands
|
||||
</a>
|
||||
</p>
|
||||
</section>
|
||||
|
||||
<!-- Using Skills Section -->
|
||||
<section id="skills">
|
||||
<h2><i class="fa-solid fa-wand-magic-sparkles"></i> Using Skills</h2>
|
||||
|
||||
<p>
|
||||
Here's the difference: agents are <em>who</em> does the work, skills are <em>what they know</em>. When you invoke a skill, you're loading a reference library into Claude's context—patterns, templates, examples, workflows. It's like handing Claude a technical manual.
|
||||
</p>
|
||||
|
||||
<h3>Invoking Skills</h3>
|
||||
<div class="card-code-block">
|
||||
<pre><code># In your prompt, reference the skill
|
||||
skill: gemini-imagegen
|
||||
|
||||
# Or ask Claude to use it
|
||||
"Use the dhh-ruby-style skill to refactor this code"</code></pre>
|
||||
</div>
|
||||
|
||||
<h3>Skill Structure</h3>
|
||||
<p>Peek inside a skill directory and you'll usually find:</p>
|
||||
<ul>
|
||||
<li><strong>SKILL.md</strong> - The main instructions (what Claude reads first)</li>
|
||||
<li><strong>references/</strong> - Deep dives on concepts and patterns</li>
|
||||
<li><strong>templates/</strong> - Copy-paste code snippets</li>
|
||||
<li><strong>workflows/</strong> - Step-by-step "how to" guides</li>
|
||||
<li><strong>scripts/</strong> - Actual executable code (when words aren't enough)</li>
|
||||
</ul>
|
||||
|
||||
<p>
|
||||
<a href="skills.html" class="button secondary">
|
||||
<i class="fa-solid fa-arrow-right"></i> View All Skills
|
||||
</a>
|
||||
</p>
|
||||
</section>
|
||||
|
||||
<!-- Code Review Workflow Guide -->
|
||||
<section id="code-review">
|
||||
<h2><i class="fa-solid fa-code-pull-request"></i> Code Review Workflow Guide</h2>
|
||||
|
||||
<p>
|
||||
You'll spend most of your time here. This workflow is why the plugin exists—to turn code review from a bottleneck into a superpower.
|
||||
</p>
|
||||
|
||||
<h3>Basic Review</h3>
|
||||
<div class="card-code-block">
|
||||
<pre><code># Review a PR
|
||||
/review 123
|
||||
|
||||
# Review current branch
|
||||
/review</code></pre>
|
||||
</div>
|
||||
|
||||
<h3>Understanding Findings</h3>
|
||||
<p>Every finding gets a priority label. Here's what they mean:</p>
|
||||
<ul>
|
||||
<li><span class="badge badge-critical">P1 Critical</span> - Don't merge until this is fixed. Think: SQL injection, data loss, crashes in production.</li>
|
||||
<li><span class="badge badge-important">P2 Important</span> - Fix before shipping. Performance regressions, N+1 queries, shaky architecture.</li>
|
||||
<li><span class="badge badge-nice">P3 Nice-to-Have</span> - Would be better, but ship without it if you need to. Documentation, minor cleanup, style issues.</li>
|
||||
</ul>
|
||||
|
||||
<h3>Working with Todo Files</h3>
|
||||
<p>After a review, you'll have a <code>todos/</code> directory full of markdown files. Each one is a single issue to fix:</p>
|
||||
<div class="card-code-block">
|
||||
<pre><code># List all pending todos
|
||||
ls todos/*-pending-*.md
|
||||
|
||||
# Triage findings
|
||||
/triage
|
||||
|
||||
# Resolve todos in parallel
|
||||
/resolve_todo_parallel</code></pre>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Creating Custom Agents -->
|
||||
<section id="creating-agents">
|
||||
<h2><i class="fa-solid fa-plus"></i> Creating Custom Agents</h2>
|
||||
|
||||
<p>
|
||||
The built-in agents cover a lot of ground, but every team has unique needs. Maybe you want a "rails-api-reviewer" that enforces your company's API standards. That's 10 minutes of work.
|
||||
</p>
|
||||
|
||||
<h3>Agent File Structure</h3>
|
||||
<div class="card-code-block">
|
||||
<pre><code>---
|
||||
name: my-custom-agent
|
||||
description: Brief description of what this agent does
|
||||
---
|
||||
|
||||
# Agent Instructions
|
||||
|
||||
You are [role description].
|
||||
|
||||
## Your Responsibilities
|
||||
1. First responsibility
|
||||
2. Second responsibility
|
||||
|
||||
## Guidelines
|
||||
- Guideline one
|
||||
- Guideline two</code></pre>
|
||||
</div>
|
||||
|
||||
<h3>Agent Location</h3>
|
||||
<p>Drop your agent file in one of these directories:</p>
|
||||
<ul>
|
||||
<li><code>.claude/agents/</code> - Just for this project (committed to git)</li>
|
||||
<li><code>~/.claude/agents/</code> - Available in all your projects (stays on your machine)</li>
|
||||
</ul>
|
||||
|
||||
<div class="callout callout-tip">
|
||||
<div class="callout-icon"><i class="fa-solid fa-lightbulb"></i></div>
|
||||
<div class="callout-content">
|
||||
<h4>The Easy Way</h4>
|
||||
<p>
|
||||
Don't write the YAML by hand. Just run <code>/create-agent-skill</code> and answer a few questions. The command generates the file, validates the format, and puts it in the right place.
|
||||
</p>
|
||||
</div>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Creating Custom Skills -->
|
||||
<section id="creating-skills">
|
||||
<h2><i class="fa-solid fa-plus"></i> Creating Custom Skills</h2>
|
||||
|
||||
<p>
|
||||
Skills are heavier than agents—they're knowledge bases, not just prompts. You're building a mini library that Claude can reference. Worth the effort for things you do repeatedly.
|
||||
</p>
|
||||
|
||||
<h3>Skill Directory Structure</h3>
|
||||
<div class="card-code-block">
|
||||
<pre><code>my-skill/
|
||||
SKILL.md # Main skill file (required)
|
||||
references/ # Supporting documentation
|
||||
concept-one.md
|
||||
concept-two.md
|
||||
templates/ # Code templates
|
||||
basic-template.md
|
||||
workflows/ # Step-by-step procedures
|
||||
workflow-one.md
|
||||
scripts/ # Executable scripts
|
||||
helper.py</code></pre>
|
||||
</div>
|
||||
|
||||
<h3>SKILL.md Format</h3>
|
||||
<div class="card-code-block">
|
||||
<pre><code>---
|
||||
name: my-skill
|
||||
description: Brief description shown when skill is invoked
|
||||
---
|
||||
|
||||
# Skill Title
|
||||
|
||||
Detailed instructions for using this skill.
|
||||
|
||||
## Quick Start
|
||||
...
|
||||
|
||||
## Reference Materials
|
||||
The skill includes references in the `references/` directory.
|
||||
|
||||
## Templates
|
||||
Use templates from the `templates/` directory.</code></pre>
|
||||
</div>
|
||||
|
||||
<div class="callout callout-tip">
|
||||
<div class="callout-icon"><i class="fa-solid fa-lightbulb"></i></div>
|
||||
<div class="callout-content">
|
||||
<h4>Get Help Building Skills</h4>
|
||||
<p>
|
||||
Type <code>skill: create-agent-skills</code> and Claude loads expert guidance on skill architecture, best practices, file organization, and validation. It's like having a senior engineer walk you through it.
|
||||
</p>
|
||||
</div>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Navigation -->
|
||||
<nav class="docs-nav-footer">
|
||||
<a href="../index.html" class="nav-prev">
|
||||
<span class="nav-label">Previous</span>
|
||||
<span class="nav-title"><i class="fa-solid fa-arrow-left"></i> Home</span>
|
||||
</a>
|
||||
<a href="agents.html" class="nav-next">
|
||||
<span class="nav-label">Next</span>
|
||||
<span class="nav-title">Agent Reference <i class="fa-solid fa-arrow-right"></i></span>
|
||||
</a>
|
||||
</nav>
|
||||
</article>
|
||||
</main>
|
||||
</div>
|
||||
|
||||
<script>
|
||||
// Sidebar toggle for mobile
|
||||
document.querySelector('[data-sidebar-toggle]')?.addEventListener('click', () => {
|
||||
document.querySelector('.docs-sidebar').classList.toggle('open');
|
||||
});
|
||||
|
||||
// Active link highlighting
|
||||
const sections = document.querySelectorAll('section[id]');
|
||||
const navLinks = document.querySelectorAll('.sidebar-nav a');
|
||||
|
||||
window.addEventListener('scroll', () => {
|
||||
let current = '';
|
||||
sections.forEach(section => {
|
||||
const sectionTop = section.offsetTop;
|
||||
if (pageYOffset >= sectionTop - 100) {
|
||||
current = section.getAttribute('id');
|
||||
}
|
||||
});
|
||||
|
||||
navLinks.forEach(link => {
|
||||
link.classList.remove('active');
|
||||
if (link.getAttribute('href') === `#${current}`) {
|
||||
link.classList.add('active');
|
||||
}
|
||||
});
|
||||
});
|
||||
</script>
|
||||
</body>
|
||||
|
||||
</html>
|
||||
@@ -1,409 +0,0 @@
|
||||
<!DOCTYPE html>
|
||||
<html lang="en" class="theme-dark">
|
||||
|
||||
<head>
|
||||
<meta charset="utf-8" />
|
||||
<title>MCP Servers Reference - Compounding Engineering</title>
|
||||
<meta content="Complete reference for the two MCP servers in the Compounding Engineering plugin." name="description" />
|
||||
<meta content="width=device-width, initial-scale=1" name="viewport" />
|
||||
|
||||
<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/font-awesome/6.5.1/css/all.min.css" />
|
||||
<link href="../css/style.css" rel="stylesheet" type="text/css" />
|
||||
<link href="../css/docs.css" rel="stylesheet" type="text/css" />
|
||||
<script src="../js/main.js" type="text/javascript" defer></script>
|
||||
</head>
|
||||
|
||||
<body>
|
||||
<div class="background-gradient"></div>
|
||||
<div class="docs-layout">
|
||||
<aside class="docs-sidebar">
|
||||
<div class="sidebar-header">
|
||||
<a href="../index.html" class="nav-brand">
|
||||
<span class="logo-icon"><i class="fa-solid fa-layer-group"></i></span>
|
||||
<span class="logo-text">CE Docs</span>
|
||||
</a>
|
||||
</div>
|
||||
<nav class="sidebar-nav">
|
||||
<div class="nav-section">
|
||||
<h3>Getting Started</h3>
|
||||
<ul>
|
||||
<li><a href="getting-started.html">Installation</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="nav-section">
|
||||
<h3>Reference</h3>
|
||||
<ul>
|
||||
<li><a href="agents.html">Agents (23)</a></li>
|
||||
<li><a href="commands.html">Commands (13)</a></li>
|
||||
<li><a href="skills.html">Skills (11)</a></li>
|
||||
<li><a href="mcp-servers.html" class="active">MCP Servers (two)</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="nav-section">
|
||||
<h3>Resources</h3>
|
||||
<ul>
|
||||
<li><a href="changelog.html">Changelog</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="nav-section">
|
||||
<h3>On This Page</h3>
|
||||
<ul>
|
||||
<li><a href="#playwright">Playwright</a></li>
|
||||
<li><a href="#context7">Context7</a></li>
|
||||
<li><a href="#manual-config">Manual Configuration</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
</nav>
|
||||
</aside>
|
||||
|
||||
<main class="docs-content">
|
||||
<div class="docs-header">
|
||||
<nav class="breadcrumb">
|
||||
<a href="../index.html">Home</a>
|
||||
<span>/</span>
|
||||
<a href="getting-started.html">Docs</a>
|
||||
<span>/</span>
|
||||
<span>MCP Servers</span>
|
||||
</nav>
|
||||
<button class="mobile-menu-toggle" data-sidebar-toggle>
|
||||
<i class="fa-solid fa-bars"></i>
|
||||
</button>
|
||||
</div>
|
||||
|
||||
<article class="docs-article">
|
||||
<h1><i class="fa-solid fa-server color-accent"></i> MCP Servers Reference</h1>
|
||||
<p class="lead">
|
||||
Think of MCP servers as power tools that plug into Claude Code. Want Claude to actually <em>open a browser</em> and click around your app? That's Playwright. Need the latest Rails docs without leaving your terminal? That's Context7. The plugin bundles both servers—they just work when you install.
|
||||
</p>
|
||||
|
||||
<div class="callout callout-warning">
|
||||
<div class="callout-icon"><i class="fa-solid fa-triangle-exclamation"></i></div>
|
||||
<div class="callout-content">
|
||||
<h4>Known Issue: Auto-Loading</h4>
|
||||
<p>
|
||||
Sometimes MCP servers don't wake up automatically. If Claude can't take screenshots or look up docs, you'll need to add them manually. See <a href="#manual-config">Manual Configuration</a> for the fix.
|
||||
</p>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Playwright -->
|
||||
<section id="playwright">
|
||||
<h2><i class="fa-brands fa-chrome"></i> Playwright</h2>
|
||||
<p>
|
||||
You know how you can tell a junior developer "open Chrome and click the login button"? Now you can tell Claude the same thing. Playwright gives Claude hands to control a real browser—clicking buttons, filling forms, taking screenshots, running JavaScript. It's like pair programming with someone who has a browser open next to you.
|
||||
</p>
|
||||
|
||||
<h3>Tools Provided</h3>
|
||||
<table class="docs-table">
|
||||
<thead>
|
||||
<tr>
|
||||
<th>Tool</th>
|
||||
<th>Description</th>
|
||||
</tr>
|
||||
</thead>
|
||||
<tbody>
|
||||
<tr>
|
||||
<td><code>browser_navigate</code></td>
|
||||
<td>Go to any URL—your localhost dev server, production, staging, that competitor's site you're studying</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><code>browser_take_screenshot</code></td>
|
||||
<td>Capture what you're seeing right now. Perfect for "does this look right?" design reviews</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><code>browser_click</code></td>
|
||||
<td>Click buttons, links, whatever. Claude finds it by text or CSS selector, just like you would</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><code>browser_fill_form</code></td>
|
||||
<td>Type into forms faster than you can. Great for testing signup flows without manual clicking</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><code>browser_snapshot</code></td>
|
||||
<td>Get the page's accessibility tree—how screen readers see it. Useful for understanding structure without HTML noise</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><code>browser_evaluate</code></td>
|
||||
<td>Run any JavaScript in the page. Check localStorage, trigger functions, read variables—full console access</td>
|
||||
</tr>
|
||||
</tbody>
|
||||
</table>
|
||||
|
||||
<h3>When You'll Use This</h3>
|
||||
<ul>
|
||||
<li><strong>Design reviews without leaving the terminal</strong> - "Take a screenshot of the new navbar on mobile" gets you a PNG in seconds</li>
|
||||
<li><strong>Testing signup flows while you code</strong> - "Fill in the registration form with test@example.com and click submit" runs the test for you</li>
|
||||
<li><strong>Debugging production issues</strong> - "Navigate to the error page and show me what's in localStorage" gives you the state without opening DevTools</li>
|
||||
<li><strong>Competitive research</strong> - "Go to competitor.com and screenshot their pricing page" builds your swipe file automatically</li>
|
||||
</ul>
|
||||
|
||||
<h3>Example Usage</h3>
|
||||
<div class="card-code-block">
|
||||
<pre><code># Just talk to Claude naturally—it knows when to use Playwright
|
||||
|
||||
# Design review
|
||||
"Take a screenshot of the login page"
|
||||
|
||||
# Testing a form
|
||||
"Navigate to /signup and fill in the email field with test@example.com"
|
||||
|
||||
# Debug JavaScript state
|
||||
"Go to localhost:3000 and run console.log(window.currentUser)"
|
||||
|
||||
# The browser runs in the background. You'll get results without switching windows.</code></pre>
|
||||
</div>
|
||||
|
||||
<h3>Configuration</h3>
|
||||
<div class="card-code-block">
|
||||
<pre><code>{
|
||||
"playwright": {
|
||||
"type": "stdio",
|
||||
"command": "npx",
|
||||
"args": ["-y", "@playwright/mcp@latest"],
|
||||
"env": {}
|
||||
}
|
||||
}</code></pre>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Context7 -->
|
||||
<section id="context7">
|
||||
<h2><i class="fa-solid fa-book-open"></i> Context7</h2>
|
||||
<p>
|
||||
Ever ask Claude about a framework and get an answer from 2023? Context7 fixes that. It's a documentation service that keeps Claude current with 100+ frameworks—Rails, React, Next.js, Django, whatever you're using. Think of it as having the official docs piped directly into Claude's brain.
|
||||
</p>
|
||||
|
||||
<h3>Tools Provided</h3>
|
||||
<table class="docs-table">
|
||||
<thead>
|
||||
<tr>
|
||||
<th>Tool</th>
|
||||
<th>Description</th>
|
||||
</tr>
|
||||
</thead>
|
||||
<tbody>
|
||||
<tr>
|
||||
<td><code>resolve-library-id</code></td>
|
||||
<td>Maps "Rails" to the actual library identifier Context7 uses. You don't call this—Claude does it automatically</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><code>get-library-docs</code></td>
|
||||
<td>Fetches the actual documentation pages. Ask "How does useEffect work?" and this grabs the latest React docs</td>
|
||||
</tr>
|
||||
</tbody>
|
||||
</table>
|
||||
|
||||
<h3>What's Covered</h3>
|
||||
<p>Over 100 frameworks and libraries. Here's a taste of what you can look up:</p>
|
||||
<div class="framework-grid">
|
||||
<div class="framework-category">
|
||||
<h4>Backend</h4>
|
||||
<ul>
|
||||
<li>Ruby on Rails</li>
|
||||
<li>Django</li>
|
||||
<li>Laravel</li>
|
||||
<li>Express</li>
|
||||
<li>FastAPI</li>
|
||||
<li>Spring Boot</li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="framework-category">
|
||||
<h4>Frontend</h4>
|
||||
<ul>
|
||||
<li>React</li>
|
||||
<li>Vue.js</li>
|
||||
<li>Angular</li>
|
||||
<li>Svelte</li>
|
||||
<li>Next.js</li>
|
||||
<li>Nuxt</li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="framework-category">
|
||||
<h4>Mobile</h4>
|
||||
<ul>
|
||||
<li>React Native</li>
|
||||
<li>Flutter</li>
|
||||
<li>SwiftUI</li>
|
||||
<li>Kotlin</li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="framework-category">
|
||||
<h4>Tools & Libraries</h4>
|
||||
<ul>
|
||||
<li>Tailwind CSS</li>
|
||||
<li>PostgreSQL</li>
|
||||
<li>Redis</li>
|
||||
<li>GraphQL</li>
|
||||
<li>Prisma</li>
|
||||
<li>And many more...</li>
|
||||
</ul>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<h3>Example Usage</h3>
|
||||
<div class="card-code-block">
|
||||
<pre><code># Just ask about the framework—Claude fetches current docs automatically
|
||||
|
||||
"Look up the Rails ActionCable documentation"
|
||||
|
||||
"How does the useEffect hook work in React?"
|
||||
|
||||
"What are the best practices for PostgreSQL indexes?"
|
||||
|
||||
# You get answers based on the latest docs, not Claude's training cutoff</code></pre>
|
||||
</div>
|
||||
|
||||
<h3>Configuration</h3>
|
||||
<div class="card-code-block">
|
||||
<pre><code>{
|
||||
"context7": {
|
||||
"type": "http",
|
||||
"url": "https://mcp.context7.com/mcp"
|
||||
}
|
||||
}</code></pre>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Manual Configuration -->
|
||||
<section id="manual-config">
|
||||
<h2><i class="fa-solid fa-gear"></i> Manual Configuration</h2>
|
||||
<p>
|
||||
If the servers don't load automatically (you'll know because Claude can't take screenshots or fetch docs), you need to wire them up yourself. It's a two-minute copy-paste job.
|
||||
</p>
|
||||
|
||||
<h3>Project-Level Configuration</h3>
|
||||
<p>To enable for just this project, add this to <code>.claude/settings.json</code> in your project root:</p>
|
||||
<div class="card-code-block">
|
||||
<pre><code>{
|
||||
"mcpServers": {
|
||||
"playwright": {
|
||||
"type": "stdio",
|
||||
"command": "npx",
|
||||
"args": ["-y", "@playwright/mcp@latest"],
|
||||
"env": {}
|
||||
},
|
||||
"context7": {
|
||||
"type": "http",
|
||||
"url": "https://mcp.context7.com/mcp"
|
||||
}
|
||||
}
|
||||
}</code></pre>
|
||||
</div>
|
||||
|
||||
<h3>Global Configuration</h3>
|
||||
<p>Or enable everywhere—every project on your machine gets these servers. Add to <code>~/.claude/settings.json</code>:</p>
|
||||
<div class="card-code-block">
|
||||
<pre><code>{
|
||||
"mcpServers": {
|
||||
"playwright": {
|
||||
"type": "stdio",
|
||||
"command": "npx",
|
||||
"args": ["-y", "@playwright/mcp@latest"],
|
||||
"env": {}
|
||||
},
|
||||
"context7": {
|
||||
"type": "http",
|
||||
"url": "https://mcp.context7.com/mcp"
|
||||
}
|
||||
}
|
||||
}</code></pre>
|
||||
</div>
|
||||
|
||||
<h3>Requirements</h3>
|
||||
<table class="docs-table">
|
||||
<thead>
|
||||
<tr>
|
||||
<th>Server</th>
|
||||
<th>Requirement</th>
|
||||
</tr>
|
||||
</thead>
|
||||
<tbody>
|
||||
<tr>
|
||||
<td>Playwright</td>
|
||||
<td>Node.js 18+ and npx</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td>Context7</td>
|
||||
<td>Internet connection (HTTP endpoint)</td>
|
||||
</tr>
|
||||
</tbody>
|
||||
</table>
|
||||
|
||||
<h3>Verifying MCP Servers</h3>
|
||||
<p>After you add the config, restart Claude Code. Then test that everything works:</p>
|
||||
<div class="card-code-block">
|
||||
<pre><code># Ask Claude what it has
|
||||
"What MCP tools do you have access to?"
|
||||
|
||||
# Test Playwright (should work now)
|
||||
"Take a screenshot of the current directory listing"
|
||||
|
||||
# Test Context7 (should fetch real docs)
|
||||
"Look up Rails Active Record documentation"
|
||||
|
||||
# If either fails, double-check your JSON syntax and file paths</code></pre>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Navigation -->
|
||||
<nav class="docs-nav-footer">
|
||||
<a href="skills.html" class="nav-prev">
|
||||
<span class="nav-label">Previous</span>
|
||||
<span class="nav-title"><i class="fa-solid fa-arrow-left"></i> Skills</span>
|
||||
</a>
|
||||
<a href="getting-started.html" class="nav-next">
|
||||
<span class="nav-label">Back to</span>
|
||||
<span class="nav-title">Getting Started <i class="fa-solid fa-arrow-right"></i></span>
|
||||
</a>
|
||||
</nav>
|
||||
</article>
|
||||
</main>
|
||||
</div>
|
||||
|
||||
<style>
|
||||
.framework-grid {
|
||||
display: grid;
|
||||
grid-template-columns: repeat(2, 1fr);
|
||||
gap: var(--space-l);
|
||||
margin: var(--space-l) 0;
|
||||
}
|
||||
|
||||
@media (min-width: 768px) {
|
||||
.framework-grid {
|
||||
grid-template-columns: repeat(4, 1fr);
|
||||
}
|
||||
}
|
||||
|
||||
.framework-category {
|
||||
background-color: var(--color-surface);
|
||||
padding: var(--space-l);
|
||||
border-radius: var(--radius-m);
|
||||
border: 1px solid var(--color-border);
|
||||
}
|
||||
|
||||
.framework-category h4 {
|
||||
margin: 0 0 var(--space-s) 0;
|
||||
color: var(--color-accent);
|
||||
font-size: var(--font-size-s);
|
||||
}
|
||||
|
||||
.framework-category ul {
|
||||
margin: 0;
|
||||
padding-left: var(--space-l);
|
||||
}
|
||||
|
||||
.framework-category li {
|
||||
margin: var(--space-xs) 0;
|
||||
font-size: var(--font-size-s);
|
||||
color: var(--color-text-secondary);
|
||||
}
|
||||
</style>
|
||||
|
||||
<script>
|
||||
document.querySelector('[data-sidebar-toggle]')?.addEventListener('click', () => {
|
||||
document.querySelector('.docs-sidebar').classList.toggle('open');
|
||||
});
|
||||
</script>
|
||||
</body>
|
||||
|
||||
</html>
|
||||
@@ -1,611 +0,0 @@
|
||||
<!DOCTYPE html>
|
||||
<html lang="en" class="theme-dark">
|
||||
|
||||
<head>
|
||||
<meta charset="utf-8" />
|
||||
<title>Skill Reference - Compounding Engineering</title>
|
||||
<meta content="Complete reference for all 12 intelligent skills in the Compounding Engineering plugin." name="description" />
|
||||
<meta content="width=device-width, initial-scale=1" name="viewport" />
|
||||
|
||||
<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/font-awesome/6.5.1/css/all.min.css" />
|
||||
<link href="../css/style.css" rel="stylesheet" type="text/css" />
|
||||
<link href="../css/docs.css" rel="stylesheet" type="text/css" />
|
||||
<script src="../js/main.js" type="text/javascript" defer></script>
|
||||
</head>
|
||||
|
||||
<body>
|
||||
<div class="background-gradient"></div>
|
||||
<div class="docs-layout">
|
||||
<aside class="docs-sidebar">
|
||||
<div class="sidebar-header">
|
||||
<a href="../index.html" class="nav-brand">
|
||||
<span class="logo-icon"><i class="fa-solid fa-layer-group"></i></span>
|
||||
<span class="logo-text">CE Docs</span>
|
||||
</a>
|
||||
</div>
|
||||
<nav class="sidebar-nav">
|
||||
<div class="nav-section">
|
||||
<h3>Getting Started</h3>
|
||||
<ul>
|
||||
<li><a href="getting-started.html">Installation</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="nav-section">
|
||||
<h3>Reference</h3>
|
||||
<ul>
|
||||
<li><a href="agents.html">Agents (27)</a></li>
|
||||
<li><a href="commands.html">Commands (19)</a></li>
|
||||
<li><a href="skills.html" class="active">Skills (12)</a></li>
|
||||
<li><a href="mcp-servers.html">MCP Servers (2)</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="nav-section">
|
||||
<h3>Resources</h3>
|
||||
<ul>
|
||||
<li><a href="changelog.html">Changelog</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="nav-section">
|
||||
<h3>On This Page</h3>
|
||||
<ul>
|
||||
<li><a href="#development-tools">Development (8)</a></li>
|
||||
<li><a href="#content-workflow">Content & Workflow (3)</a></li>
|
||||
<li><a href="#image-generation">Image Generation (1)</a></li>
|
||||
</ul>
|
||||
</div>
|
||||
</nav>
|
||||
</aside>
|
||||
|
||||
<main class="docs-content">
|
||||
<div class="docs-header">
|
||||
<nav class="breadcrumb">
|
||||
<a href="../index.html">Home</a>
|
||||
<span>/</span>
|
||||
<a href="getting-started.html">Docs</a>
|
||||
<span>/</span>
|
||||
<span>Skills</span>
|
||||
</nav>
|
||||
<button class="mobile-menu-toggle" data-sidebar-toggle>
|
||||
<i class="fa-solid fa-bars"></i>
|
||||
</button>
|
||||
</div>
|
||||
|
||||
<article class="docs-article">
|
||||
<h1><i class="fa-solid fa-wand-magic-sparkles color-accent"></i> Skill Reference</h1>
|
||||
<p class="lead">
|
||||
Think of skills as reference manuals that Claude Code can read mid-conversation. When you're writing Rails code and want DHH's style, or building a gem like Andrew Kane would, you don't need to paste documentation—just invoke the skill. Claude reads it, absorbs the patterns, and writes code that way.
|
||||
</p>
|
||||
|
||||
<div class="usage-box">
|
||||
<h3>How to Use Skills</h3>
|
||||
<div class="card-code-block">
|
||||
<pre><code># In your prompt, reference the skill
|
||||
skill: [skill-name]
|
||||
|
||||
# Examples
|
||||
skill: gemini-imagegen
|
||||
skill: dhh-rails-style
|
||||
skill: create-agent-skills</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="callout callout-info">
|
||||
<div class="callout-icon"><i class="fa-solid fa-circle-info"></i></div>
|
||||
<div class="callout-content">
|
||||
<h4>Skills vs Agents</h4>
|
||||
<p>
|
||||
<strong>Agents</strong> are personas—they <em>do</em> things. <strong>Skills</strong> are knowledge—they teach Claude <em>how</em> to do things. Use <code>claude agent [name]</code> when you want someone to review your code. Use <code>skill: [name]</code> when you want to write code in a particular style yourself.
|
||||
</p>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Development Tools -->
|
||||
<section id="development-tools">
|
||||
<h2><i class="fa-solid fa-code"></i> Development Tools (8)</h2>
|
||||
<p>These skills teach Claude specific coding styles and architectural patterns. Use them when you want code that follows a particular philosophy—not just any working code, but code that looks like it was written by a specific person or framework.</p>
|
||||
|
||||
<div class="skill-detail" id="create-agent-skills">
|
||||
<div class="skill-detail-header">
|
||||
<h3>create-agent-skills</h3>
|
||||
<span class="skill-badge">Meta</span>
|
||||
</div>
|
||||
<p class="skill-detail-description">
|
||||
You're writing a skill right now, but you're not sure if you're structuring the SKILL.md file correctly. Should the examples go before the theory? How do you organize workflows vs. references? This skill is the answer—it's the master template for building skills themselves.
|
||||
</p>
|
||||
<h4>Capabilities</h4>
|
||||
<ul>
|
||||
<li>Skill architecture and best practices</li>
|
||||
<li>Router pattern for complex multi-step skills</li>
|
||||
<li>Progressive disclosure design principles</li>
|
||||
<li>SKILL.md structure guidance</li>
|
||||
<li>Asset management (workflows, references, templates, scripts)</li>
|
||||
<li>XML structure patterns</li>
|
||||
</ul>
|
||||
<h4>Workflows Included</h4>
|
||||
<ul>
|
||||
<li><code>create-new-skill</code> - Start from scratch</li>
|
||||
<li><code>add-reference</code> - Add reference documentation</li>
|
||||
<li><code>add-template</code> - Add code templates</li>
|
||||
<li><code>add-workflow</code> - Add step-by-step procedures</li>
|
||||
<li><code>add-script</code> - Add executable scripts</li>
|
||||
<li><code>audit-skill</code> - Validate skill structure</li>
|
||||
<li><code>verify-skill</code> - Test skill functionality</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>skill: create-agent-skills</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="skill-detail" id="skill-creator">
|
||||
<div class="skill-detail-header">
|
||||
<h3>skill-creator</h3>
|
||||
<span class="skill-badge">Meta</span>
|
||||
</div>
|
||||
<p class="skill-detail-description">
|
||||
The simpler, step-by-step version of <code>create-agent-skills</code>. When you just want a checklist to follow from blank file to packaged skill, use this. It's less about theory, more about "do step 1, then step 2."
|
||||
</p>
|
||||
<h4>6-Step Process</h4>
|
||||
<ol>
|
||||
<li>Understand skill usage patterns with examples</li>
|
||||
<li>Plan reusable skill contents</li>
|
||||
<li>Initialize skill using template</li>
|
||||
<li>Edit skill with clear instructions</li>
|
||||
<li>Package skill into distributable zip</li>
|
||||
<li>Iterate based on testing feedback</li>
|
||||
</ol>
|
||||
<div class="card-code-block">
|
||||
<pre><code>skill: skill-creator</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="skill-detail" id="dhh-rails-style">
|
||||
<div class="skill-detail-header">
|
||||
<h3>dhh-rails-style</h3>
|
||||
<span class="skill-badge">Rails</span>
|
||||
</div>
|
||||
<p class="skill-detail-description">
|
||||
Comprehensive 37signals Rails conventions based on Marc Köhlbrugge's analysis of 265 PRs from the Fizzy codebase. Covers everything from REST mapping to state-as-records, Turbo/Stimulus patterns, CSS with OKLCH colors, Minitest with fixtures, and Solid Queue/Cache/Cable patterns.
|
||||
</p>
|
||||
<h4>Key Patterns</h4>
|
||||
<ul>
|
||||
<li><strong>REST Purity</strong> - Verbs become nouns (close → closure)</li>
|
||||
<li><strong>State as Records</strong> - Boolean columns → separate records</li>
|
||||
<li><strong>Fat Models</strong> - Business logic, authorization, broadcasting</li>
|
||||
<li><strong>Thin Controllers</strong> - 1-5 line actions with concerns</li>
|
||||
<li><strong>Current Attributes</strong> - Request context everywhere</li>
|
||||
<li><strong>Hotwire/Turbo</strong> - Model-level broadcasting, morphing</li>
|
||||
</ul>
|
||||
<h4>Reference Files (6)</h4>
|
||||
<ul>
|
||||
<li><code>controllers.md</code> - REST mapping, concerns, Turbo responses</li>
|
||||
<li><code>models.md</code> - Concerns, state records, callbacks, POROs</li>
|
||||
<li><code>frontend.md</code> - Turbo, Stimulus, CSS layers, OKLCH</li>
|
||||
<li><code>architecture.md</code> - Routing, auth, jobs, caching</li>
|
||||
<li><code>testing.md</code> - Minitest, fixtures, integration tests</li>
|
||||
<li><code>gems.md</code> - What to use vs avoid, decision framework</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>skill: dhh-rails-style</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="skill-detail" id="andrew-kane-gem-writer">
|
||||
<div class="skill-detail-header">
|
||||
<h3>andrew-kane-gem-writer</h3>
|
||||
<span class="skill-badge">Ruby</span>
|
||||
</div>
|
||||
<p class="skill-detail-description">
|
||||
Andrew Kane has written 100+ Ruby gems with 374 million downloads. Every gem follows the same patterns: minimal dependencies, class macro DSLs, Rails integration without Rails coupling. When you're building a gem and want it to feel production-ready from day one, this is how you do it.
|
||||
</p>
|
||||
<h4>Philosophy</h4>
|
||||
<ul>
|
||||
<li>Simplicity over cleverness</li>
|
||||
<li>Zero or minimal dependencies</li>
|
||||
<li>Explicit code over metaprogramming</li>
|
||||
<li>Rails integration without Rails coupling</li>
|
||||
</ul>
|
||||
<h4>Key Patterns</h4>
|
||||
<ul>
|
||||
<li>Class macro DSL for configuration</li>
|
||||
<li><code>ActiveSupport.on_load</code> for Rails integration</li>
|
||||
<li><code>class << self</code> with <code>attr_accessor</code></li>
|
||||
<li>Railtie pattern for hooks</li>
|
||||
<li>Minitest (no RSpec)</li>
|
||||
</ul>
|
||||
<h4>Reference Files</h4>
|
||||
<ul>
|
||||
<li><code>references/module-organization.md</code></li>
|
||||
<li><code>references/rails-integration.md</code></li>
|
||||
<li><code>references/database-adapters.md</code></li>
|
||||
<li><code>references/testing-patterns.md</code></li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>skill: andrew-kane-gem-writer</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="skill-detail" id="dspy-ruby">
|
||||
<div class="skill-detail-header">
|
||||
<h3>dspy-ruby</h3>
|
||||
<span class="skill-badge">AI</span>
|
||||
</div>
|
||||
<p class="skill-detail-description">
|
||||
You're adding AI features to your Rails app, but you don't want brittle prompt strings scattered everywhere. DSPy.rb gives you type-safe signatures, composable predictors, and tool-using agents. This skill shows you how to use it—from basic inference to ReAct agents that iterate until they get the answer right.
|
||||
</p>
|
||||
<h4>Predictor Types</h4>
|
||||
<ul>
|
||||
<li><strong>Predict</strong> - Basic inference</li>
|
||||
<li><strong>ChainOfThought</strong> - Reasoning with explanations</li>
|
||||
<li><strong>ReAct</strong> - Tool-using agents with iteration</li>
|
||||
<li><strong>CodeAct</strong> - Dynamic code generation</li>
|
||||
</ul>
|
||||
<h4>Supported Providers</h4>
|
||||
<ul>
|
||||
<li>OpenAI (GPT-4, GPT-4o-mini)</li>
|
||||
<li>Anthropic Claude</li>
|
||||
<li>Google Gemini</li>
|
||||
<li>Ollama (free, local)</li>
|
||||
<li>OpenRouter</li>
|
||||
</ul>
|
||||
<h4>Requirements</h4>
|
||||
<table class="docs-table">
|
||||
<tr>
|
||||
<td><code>OPENAI_API_KEY</code></td>
|
||||
<td>For OpenAI provider</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><code>ANTHROPIC_API_KEY</code></td>
|
||||
<td>For Anthropic provider</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><code>GOOGLE_API_KEY</code></td>
|
||||
<td>For Gemini provider</td>
|
||||
</tr>
|
||||
</table>
|
||||
<div class="card-code-block">
|
||||
<pre><code>skill: dspy-ruby</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="skill-detail" id="frontend-design">
|
||||
<div class="skill-detail-header">
|
||||
<h3>frontend-design</h3>
|
||||
<span class="skill-badge">Design</span>
|
||||
</div>
|
||||
<p class="skill-detail-description">
|
||||
You've seen what AI usually generates: Inter font, purple gradients, rounded corners on everything. This skill teaches Claude to design interfaces that don't look like every other AI-generated site. It's about purposeful typography, unexpected color palettes, and interfaces with personality.
|
||||
</p>
|
||||
<h4>Design Thinking</h4>
|
||||
<ul>
|
||||
<li><strong>Purpose</strong> - What is the interface for?</li>
|
||||
<li><strong>Tone</strong> - What feeling should it evoke?</li>
|
||||
<li><strong>Constraints</strong> - Technical and brand limitations</li>
|
||||
<li><strong>Differentiation</strong> - How to stand out</li>
|
||||
</ul>
|
||||
<h4>Focus Areas</h4>
|
||||
<ul>
|
||||
<li>Typography with distinctive font choices</li>
|
||||
<li>Color & theme coherence with CSS variables</li>
|
||||
<li>Motion and animation patterns</li>
|
||||
<li>Spatial composition with asymmetry</li>
|
||||
<li>Backgrounds (gradients, textures, patterns)</li>
|
||||
</ul>
|
||||
<div class="callout callout-tip">
|
||||
<div class="callout-icon"><i class="fa-solid fa-lightbulb"></i></div>
|
||||
<div class="callout-content">
|
||||
<p>Avoids generic AI aesthetics like Inter fonts, purple gradients, and rounded corners everywhere.</p>
|
||||
</div>
|
||||
</div>
|
||||
<div class="card-code-block">
|
||||
<pre><code>skill: frontend-design</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="skill-detail" id="compound-docs">
|
||||
<div class="skill-detail-header">
|
||||
<h3>compound-docs</h3>
|
||||
<span class="skill-badge">Docs</span>
|
||||
</div>
|
||||
<p class="skill-detail-description">
|
||||
You just fixed a weird build error after an hour of debugging. Tomorrow you'll forget how you fixed it. This skill automatically detects when you solve something (phrases like "that worked" or "it's fixed") and documents it with YAML frontmatter so you can find it again. Each documented solution compounds your team's knowledge.
|
||||
</p>
|
||||
<h4>Auto-Triggers</h4>
|
||||
<p>Phrases: "that worked", "it's fixed", "working now", "problem solved"</p>
|
||||
<h4>7-Step Process</h4>
|
||||
<ol>
|
||||
<li>Detect confirmation phrase</li>
|
||||
<li>Gather context (module, symptom, investigation, root cause)</li>
|
||||
<li>Check existing docs for similar issues</li>
|
||||
<li>Generate filename</li>
|
||||
<li>Validate YAML frontmatter</li>
|
||||
<li>Create documentation in category directory</li>
|
||||
<li>Cross-reference related issues</li>
|
||||
</ol>
|
||||
<h4>Categories</h4>
|
||||
<ul>
|
||||
<li><code>build-errors/</code></li>
|
||||
<li><code>test-failures/</code></li>
|
||||
<li><code>runtime-errors/</code></li>
|
||||
<li><code>performance-issues/</code></li>
|
||||
<li><code>database-issues/</code></li>
|
||||
<li><code>security-issues/</code></li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>skill: compound-docs</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="skill-detail" id="agent-native-architecture">
|
||||
<div class="skill-detail-header">
|
||||
<h3>agent-native-architecture</h3>
|
||||
<span class="skill-badge">AI</span>
|
||||
</div>
|
||||
<p class="skill-detail-description">
|
||||
Build AI agents using prompt-native architecture where features are defined in prompts, not code. When creating autonomous agents, designing MCP servers, or implementing self-modifying systems, this skill guides the "trust the agent's intelligence" philosophy.
|
||||
</p>
|
||||
<h4>Key Patterns</h4>
|
||||
<ul>
|
||||
<li><strong>Prompt-Native Features</strong> - Define features in prompts, not code</li>
|
||||
<li><strong>MCP Tool Design</strong> - Build tools agents can use effectively</li>
|
||||
<li><strong>System Prompts</strong> - Write instructions that guide agent behavior</li>
|
||||
<li><strong>Self-Modification</strong> - Allow agents to improve their own prompts</li>
|
||||
</ul>
|
||||
<h4>Core Principle</h4>
|
||||
<p>Whatever the user can do, the agent can do. Whatever the user can see, the agent can see.</p>
|
||||
<div class="card-code-block">
|
||||
<pre><code>skill: agent-native-architecture</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Content & Workflow -->
|
||||
<section id="content-workflow">
|
||||
<h2><i class="fa-solid fa-pen-fancy"></i> Content & Workflow (3)</h2>
|
||||
<p>Writing, editing, and organizing work. These skills handle everything from style guide compliance to git worktree management—the meta-work that makes the real work easier.</p>
|
||||
|
||||
<div class="skill-detail" id="every-style-editor">
|
||||
<div class="skill-detail-header">
|
||||
<h3>every-style-editor</h3>
|
||||
<span class="skill-badge">Content</span>
|
||||
</div>
|
||||
<p class="skill-detail-description">
|
||||
You wrote a draft, but you're not sure if it matches Every's style guide. Should "internet" be capitalized? Is this comma splice allowed? This skill does a four-phase line-by-line review: context, detailed edits, mechanical checks, and actionable recommendations. It's like having a copy editor who never gets tired.
|
||||
</p>
|
||||
<h4>Four-Phase Review</h4>
|
||||
<ol>
|
||||
<li><strong>Initial Assessment</strong> - Context, type, audience, tone</li>
|
||||
<li><strong>Detailed Line Edit</strong> - Sentence structure, punctuation, capitalization</li>
|
||||
<li><strong>Mechanical Review</strong> - Spacing, formatting, consistency</li>
|
||||
<li><strong>Recommendations</strong> - Actionable improvement suggestions</li>
|
||||
</ol>
|
||||
<h4>Style Checks</h4>
|
||||
<ul>
|
||||
<li>Grammar and punctuation</li>
|
||||
<li>Style guide compliance</li>
|
||||
<li>Capitalization rules</li>
|
||||
<li>Word choice optimization</li>
|
||||
<li>Formatting consistency</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>skill: every-style-editor</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="skill-detail" id="file-todos">
|
||||
<div class="skill-detail-header">
|
||||
<h3>file-todos</h3>
|
||||
<span class="skill-badge">Workflow</span>
|
||||
</div>
|
||||
<p class="skill-detail-description">
|
||||
Your todo list is a bunch of markdown files in a <code>todos/</code> directory. Each filename encodes status, priority, and description. No database, no UI, just files with YAML frontmatter. When you need to track work without setting up Jira, this is the system.
|
||||
</p>
|
||||
<h4>File Format</h4>
|
||||
<div class="card-code-block">
|
||||
<pre><code># Naming convention
|
||||
{issue_id}-{status}-{priority}-{description}.md
|
||||
|
||||
# Examples
|
||||
001-pending-p1-security-vulnerability.md
|
||||
002-ready-p2-performance-optimization.md
|
||||
003-complete-p3-code-cleanup.md</code></pre>
|
||||
</div>
|
||||
<h4>Status Values</h4>
|
||||
<ul>
|
||||
<li><code>pending</code> - Needs triage</li>
|
||||
<li><code>ready</code> - Approved for work</li>
|
||||
<li><code>complete</code> - Done</li>
|
||||
</ul>
|
||||
<h4>Priority Values</h4>
|
||||
<ul>
|
||||
<li><code>p1</code> - Critical</li>
|
||||
<li><code>p2</code> - Important</li>
|
||||
<li><code>p3</code> - Nice-to-have</li>
|
||||
</ul>
|
||||
<h4>YAML Frontmatter</h4>
|
||||
<div class="card-code-block">
|
||||
<pre><code>---
|
||||
status: pending
|
||||
priority: p1
|
||||
issue_id: "001"
|
||||
tags: [security, authentication]
|
||||
dependencies: []
|
||||
---</code></pre>
|
||||
</div>
|
||||
<div class="card-code-block">
|
||||
<pre><code>skill: file-todos</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="skill-detail" id="git-worktree">
|
||||
<div class="skill-detail-header">
|
||||
<h3>git-worktree</h3>
|
||||
<span class="skill-badge">Git</span>
|
||||
</div>
|
||||
<p class="skill-detail-description">
|
||||
You're working on a feature branch, but you need to review a PR without losing your current work. Git worktrees let you have multiple branches checked out simultaneously in separate directories. This skill manages them—create, switch, cleanup—so you can context-switch without stashing or committing half-finished code.
|
||||
</p>
|
||||
<h4>Commands</h4>
|
||||
<div class="card-code-block">
|
||||
<pre><code># Create new worktree
|
||||
bash scripts/worktree-manager.sh create feature-login
|
||||
|
||||
# List worktrees
|
||||
bash scripts/worktree-manager.sh list
|
||||
|
||||
# Switch to worktree
|
||||
bash scripts/worktree-manager.sh switch feature-login
|
||||
|
||||
# Clean up completed worktrees
|
||||
bash scripts/worktree-manager.sh cleanup</code></pre>
|
||||
</div>
|
||||
<h4>Integration</h4>
|
||||
<ul>
|
||||
<li>Works with <code>/review</code> for isolated PR analysis</li>
|
||||
<li>Works with <code>/work</code> for parallel feature development</li>
|
||||
</ul>
|
||||
<h4>Requirements</h4>
|
||||
<ul>
|
||||
<li>Git 2.8+ (for worktree support)</li>
|
||||
<li>Worktrees stored in <code>.worktrees/</code> directory</li>
|
||||
</ul>
|
||||
<div class="card-code-block">
|
||||
<pre><code>skill: git-worktree</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Image Generation -->
|
||||
<section id="image-generation">
|
||||
<h2><i class="fa-solid fa-image"></i> Image Generation (1)</h2>
|
||||
<p>Generate images with AI. Not stock photos you found on Unsplash—images you describe and the model creates.</p>
|
||||
|
||||
<div class="skill-detail featured" id="gemini-imagegen">
|
||||
<div class="skill-detail-header">
|
||||
<h3>gemini-imagegen</h3>
|
||||
<span class="skill-badge highlight">AI Images</span>
|
||||
</div>
|
||||
<p class="skill-detail-description">
|
||||
Need a logo with specific text? A product mockup on a marble surface? An illustration in a kawaii style? This skill wraps Google's Gemini image generation API. You describe what you want, it generates it. You can edit existing images, refine over multiple turns, or compose from reference images. All through simple Python scripts.
|
||||
</p>
|
||||
|
||||
<h4>Features</h4>
|
||||
<div class="skill-features">
|
||||
<div class="feature-item"><i class="fa-solid fa-check"></i> Text-to-image generation</div>
|
||||
<div class="feature-item"><i class="fa-solid fa-check"></i> Image editing & manipulation</div>
|
||||
<div class="feature-item"><i class="fa-solid fa-check"></i> Multi-turn iterative refinement</div>
|
||||
<div class="feature-item"><i class="fa-solid fa-check"></i> Multiple reference images (up to 14)</div>
|
||||
<div class="feature-item"><i class="fa-solid fa-check"></i> Google Search grounding (Pro)</div>
|
||||
</div>
|
||||
|
||||
<h4>Available Models</h4>
|
||||
<table class="docs-table">
|
||||
<thead>
|
||||
<tr>
|
||||
<th>Model</th>
|
||||
<th>Resolution</th>
|
||||
<th>Best For</th>
|
||||
</tr>
|
||||
</thead>
|
||||
<tbody>
|
||||
<tr>
|
||||
<td><code>gemini-2.5-flash-image</code></td>
|
||||
<td>1024px</td>
|
||||
<td>Speed, high-volume tasks</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><code>gemini-3-pro-image-preview</code></td>
|
||||
<td>Up to 4K</td>
|
||||
<td>Professional assets, complex instructions</td>
|
||||
</tr>
|
||||
</tbody>
|
||||
</table>
|
||||
|
||||
<h4>Quick Start</h4>
|
||||
<div class="card-code-block">
|
||||
<pre><code># Text-to-image
|
||||
python scripts/generate_image.py "A cat wearing a wizard hat" output.png
|
||||
|
||||
# Edit existing image
|
||||
python scripts/edit_image.py input.png "Add a rainbow in the background" output.png
|
||||
|
||||
# Multi-turn chat
|
||||
python scripts/multi_turn_chat.py</code></pre>
|
||||
</div>
|
||||
|
||||
<h4>Image Configuration</h4>
|
||||
<div class="card-code-block">
|
||||
<pre><code>from google.genai import types
|
||||
|
||||
response = client.models.generate_content(
|
||||
model="gemini-3-pro-image-preview",
|
||||
contents=[prompt],
|
||||
config=types.GenerateContentConfig(
|
||||
response_modalities=['TEXT', 'IMAGE'],
|
||||
image_config=types.ImageConfig(
|
||||
aspect_ratio="16:9", # 1:1, 2:3, 3:2, 4:3, 16:9, 21:9
|
||||
image_size="2K" # 1K, 2K, 4K (Pro only)
|
||||
),
|
||||
)
|
||||
)</code></pre>
|
||||
</div>
|
||||
|
||||
<h4>Prompting Best Practices</h4>
|
||||
<ul>
|
||||
<li><strong>Photorealistic</strong> - Include camera details: lens type, lighting, angle, mood</li>
|
||||
<li><strong>Stylized Art</strong> - Specify style explicitly: kawaii, cel-shading, bold outlines</li>
|
||||
<li><strong>Text in Images</strong> - Be explicit about font style and placement (use Pro model)</li>
|
||||
<li><strong>Product Mockups</strong> - Describe lighting setup and surface</li>
|
||||
</ul>
|
||||
|
||||
<h4>Requirements</h4>
|
||||
<table class="docs-table">
|
||||
<tr>
|
||||
<td><code>GEMINI_API_KEY</code></td>
|
||||
<td>Required environment variable</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><code>google-genai</code></td>
|
||||
<td>Python package</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><code>pillow</code></td>
|
||||
<td>Python package for image handling</td>
|
||||
</tr>
|
||||
</table>
|
||||
|
||||
<div class="callout callout-info">
|
||||
<div class="callout-icon"><i class="fa-solid fa-circle-info"></i></div>
|
||||
<div class="callout-content">
|
||||
<p>All generated images include SynthID watermarks. Image-only mode won't work with Google Search grounding.</p>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="card-code-block">
|
||||
<pre><code>skill: gemini-imagegen</code></pre>
|
||||
</div>
|
||||
</div>
|
||||
</section>
|
||||
|
||||
<!-- Navigation -->
|
||||
<nav class="docs-nav-footer">
|
||||
<a href="commands.html" class="nav-prev">
|
||||
<span class="nav-label">Previous</span>
|
||||
<span class="nav-title"><i class="fa-solid fa-arrow-left"></i> Commands</span>
|
||||
</a>
|
||||
<a href="mcp-servers.html" class="nav-next">
|
||||
<span class="nav-label">Next</span>
|
||||
<span class="nav-title">MCP Servers <i class="fa-solid fa-arrow-right"></i></span>
|
||||
</a>
|
||||
</nav>
|
||||
</article>
|
||||
</main>
|
||||
</div>
|
||||
|
||||
<script>
|
||||
document.querySelector('[data-sidebar-toggle]')?.addEventListener('click', () => {
|
||||
document.querySelector('.docs-sidebar').classList.toggle('open');
|
||||
});
|
||||
</script>
|
||||
</body>
|
||||
|
||||
</html>
|
||||
328
docs/plans/2026-02-14-feat-add-copilot-converter-target-plan.md
Normal file
328
docs/plans/2026-02-14-feat-add-copilot-converter-target-plan.md
Normal file
@@ -0,0 +1,328 @@
|
||||
---
|
||||
title: "feat: Add GitHub Copilot converter target"
|
||||
type: feat
|
||||
date: 2026-02-14
|
||||
status: complete
|
||||
---
|
||||
|
||||
# feat: Add GitHub Copilot Converter Target
|
||||
|
||||
## Overview
|
||||
|
||||
Add GitHub Copilot as a converter target following the established `TargetHandler` pattern. This converts the compound-engineering Claude Code plugin into Copilot's native format: custom agents (`.agent.md`), agent skills (`SKILL.md`), and MCP server configuration JSON.
|
||||
|
||||
**Brainstorm:** `docs/brainstorms/2026-02-14-copilot-converter-target-brainstorm.md`
|
||||
|
||||
## Problem Statement
|
||||
|
||||
The CLI tool (`compound`) already supports converting Claude Code plugins to 5 target formats (OpenCode, Codex, Droid, Cursor, Pi). GitHub Copilot is a widely-used AI coding assistant that now supports custom agents, skills, and MCP servers — but there's no converter target for it.
|
||||
|
||||
## Proposed Solution
|
||||
|
||||
Follow the existing converter pattern exactly:
|
||||
|
||||
1. Define types (`src/types/copilot.ts`)
|
||||
2. Implement converter (`src/converters/claude-to-copilot.ts`)
|
||||
3. Implement writer (`src/targets/copilot.ts`)
|
||||
4. Register target (`src/targets/index.ts`)
|
||||
5. Add sync support (`src/sync/copilot.ts`, `src/commands/sync.ts`)
|
||||
6. Write tests and documentation
|
||||
|
||||
### Component Mapping
|
||||
|
||||
| Claude Code | Copilot | Output Path |
|
||||
|-------------|---------|-------------|
|
||||
| Agents (`.md`) | Custom Agents (`.agent.md`) | `.github/agents/{name}.agent.md` |
|
||||
| Commands (`.md`) | Agent Skills (`SKILL.md`) | `.github/skills/{name}/SKILL.md` |
|
||||
| Skills (`SKILL.md`) | Agent Skills (`SKILL.md`) | `.github/skills/{name}/SKILL.md` |
|
||||
| MCP Servers | Config JSON | `.github/copilot-mcp-config.json` |
|
||||
| Hooks | Skipped | Warning to stderr |
|
||||
|
||||
## Technical Approach
|
||||
|
||||
### Phase 1: Types
|
||||
|
||||
**File:** `src/types/copilot.ts`
|
||||
|
||||
```typescript
|
||||
export type CopilotAgent = {
|
||||
name: string
|
||||
content: string // Full .agent.md content with frontmatter
|
||||
}
|
||||
|
||||
export type CopilotGeneratedSkill = {
|
||||
name: string
|
||||
content: string // SKILL.md content with frontmatter
|
||||
}
|
||||
|
||||
export type CopilotSkillDir = {
|
||||
name: string
|
||||
sourceDir: string
|
||||
}
|
||||
|
||||
export type CopilotMcpServer = {
|
||||
type: string
|
||||
command?: string
|
||||
args?: string[]
|
||||
url?: string
|
||||
tools: string[]
|
||||
env?: Record<string, string>
|
||||
headers?: Record<string, string>
|
||||
}
|
||||
|
||||
export type CopilotBundle = {
|
||||
agents: CopilotAgent[]
|
||||
generatedSkills: CopilotGeneratedSkill[]
|
||||
skillDirs: CopilotSkillDir[]
|
||||
mcpConfig?: Record<string, CopilotMcpServer>
|
||||
}
|
||||
```
|
||||
|
||||
### Phase 2: Converter
|
||||
|
||||
**File:** `src/converters/claude-to-copilot.ts`
|
||||
|
||||
**Agent conversion:**
|
||||
- Frontmatter: `description` (required, fallback to `"Converted from Claude agent {name}"`), `tools: ["*"]`, `infer: true`
|
||||
- Pass through `model` if present
|
||||
- Fold `capabilities` into body as `## Capabilities` section (same as Cursor)
|
||||
- Use `formatFrontmatter()` utility
|
||||
- Warn if body exceeds 30,000 characters (`.length`)
|
||||
|
||||
**Command → Skill conversion:**
|
||||
- Convert to SKILL.md format with frontmatter: `name`, `description`
|
||||
- Flatten namespaced names: `workflows:plan` → `plan`
|
||||
- Drop `allowed-tools`, `model`, `disable-model-invocation` silently
|
||||
- Include `argument-hint` as `## Arguments` section in body
|
||||
|
||||
**Skill pass-through:**
|
||||
- Map to `CopilotSkillDir` as-is (same as Cursor)
|
||||
|
||||
**MCP server conversion:**
|
||||
- Transform env var names: `API_KEY` → `COPILOT_MCP_API_KEY`
|
||||
- Skip vars already prefixed with `COPILOT_MCP_`
|
||||
- Add `type: "local"` for command-based servers, `type: "sse"` for URL-based
|
||||
- Set `tools: ["*"]` for all servers
|
||||
|
||||
**Content transformation (`transformContentForCopilot`):**
|
||||
|
||||
| Pattern | Input | Output |
|
||||
|---------|-------|--------|
|
||||
| Task calls | `Task repo-research-analyst(desc)` | `Use the repo-research-analyst skill to: desc` |
|
||||
| Slash commands | `/workflows:plan` | `/plan` |
|
||||
| Path rewriting | `.claude/` | `.github/` |
|
||||
| Home path rewriting | `~/.claude/` | `~/.copilot/` |
|
||||
| Agent references | `@security-sentinel` | `the security-sentinel agent` |
|
||||
|
||||
**Hooks:** Warn to stderr if present, skip.
|
||||
|
||||
### Phase 3: Writer
|
||||
|
||||
**File:** `src/targets/copilot.ts`
|
||||
|
||||
**Path resolution:**
|
||||
- If `outputRoot` basename is `.github`, write directly into it (avoid `.github/.github/` double-nesting)
|
||||
- Otherwise, nest under `.github/`
|
||||
|
||||
**Write operations:**
|
||||
- Agents → `.github/agents/{name}.agent.md` (note: `.agent.md` extension)
|
||||
- Generated skills (from commands) → `.github/skills/{name}/SKILL.md`
|
||||
- Skill dirs → `.github/skills/{name}/` (copy via `copyDir`)
|
||||
- MCP config → `.github/copilot-mcp-config.json` (backup existing with `backupFile`)
|
||||
|
||||
### Phase 4: Target Registration
|
||||
|
||||
**File:** `src/targets/index.ts`
|
||||
|
||||
Add import and register:
|
||||
|
||||
```typescript
|
||||
import { convertClaudeToCopilot } from "../converters/claude-to-copilot"
|
||||
import { writeCopilotBundle } from "./copilot"
|
||||
|
||||
// In targets record:
|
||||
copilot: {
|
||||
name: "copilot",
|
||||
implemented: true,
|
||||
convert: convertClaudeToCopilot as TargetHandler<CopilotBundle>["convert"],
|
||||
write: writeCopilotBundle as TargetHandler<CopilotBundle>["write"],
|
||||
},
|
||||
```
|
||||
|
||||
### Phase 5: Sync Support
|
||||
|
||||
**File:** `src/sync/copilot.ts`
|
||||
|
||||
Follow the Cursor sync pattern (`src/sync/cursor.ts`):
|
||||
- Symlink skills to `.github/skills/` using `forceSymlink`
|
||||
- Validate skill names with `isValidSkillName`
|
||||
- Convert MCP servers with `COPILOT_MCP_` prefix transformation
|
||||
- Merge MCP config into existing `.github/copilot-mcp-config.json`
|
||||
|
||||
**File:** `src/commands/sync.ts`
|
||||
|
||||
- Add `"copilot"` to `validTargets` array
|
||||
- Add case in `resolveOutputRoot()`: `case "copilot": return path.join(process.cwd(), ".github")`
|
||||
- Add import and switch case for `syncToCopilot`
|
||||
- Update meta description to include "Copilot"
|
||||
|
||||
### Phase 6: Tests
|
||||
|
||||
**File:** `tests/copilot-converter.test.ts`
|
||||
|
||||
Test cases (following `tests/cursor-converter.test.ts` pattern):
|
||||
|
||||
```
|
||||
describe("convertClaudeToCopilot")
|
||||
✓ converts agents to .agent.md with Copilot frontmatter
|
||||
✓ agent description is required, fallback generated if missing
|
||||
✓ agent with empty body gets default body
|
||||
✓ agent capabilities are prepended to body
|
||||
✓ agent model field is passed through
|
||||
✓ agent tools defaults to ["*"]
|
||||
✓ agent infer defaults to true
|
||||
✓ warns when agent body exceeds 30k characters
|
||||
✓ converts commands to skills with SKILL.md format
|
||||
✓ flattens namespaced command names
|
||||
✓ command name collision after flattening is deduplicated
|
||||
✓ command allowedTools is silently dropped
|
||||
✓ command with argument-hint gets Arguments section
|
||||
✓ passes through skill directories
|
||||
✓ skill and generated skill name collision is deduplicated
|
||||
✓ converts MCP servers with COPILOT_MCP_ prefix
|
||||
✓ MCP env vars already prefixed are not double-prefixed
|
||||
✓ MCP servers get type field (local vs sse)
|
||||
✓ warns when hooks are present
|
||||
✓ no warning when hooks are absent
|
||||
✓ plugin with zero agents produces empty agents array
|
||||
✓ plugin with only skills works
|
||||
|
||||
describe("transformContentForCopilot")
|
||||
✓ rewrites .claude/ paths to .github/
|
||||
✓ rewrites ~/.claude/ paths to ~/.copilot/
|
||||
✓ transforms Task agent calls to skill references
|
||||
✓ flattens slash commands
|
||||
✓ transforms @agent references to agent references
|
||||
```
|
||||
|
||||
**File:** `tests/copilot-writer.test.ts`
|
||||
|
||||
Test cases (following `tests/cursor-writer.test.ts` pattern):
|
||||
|
||||
```
|
||||
describe("writeCopilotBundle")
|
||||
✓ writes agents, generated skills, copied skills, and MCP config
|
||||
✓ agents use .agent.md file extension
|
||||
✓ writes directly into .github output root without double-nesting
|
||||
✓ handles empty bundles gracefully
|
||||
✓ writes multiple agents as separate .agent.md files
|
||||
✓ backs up existing copilot-mcp-config.json before overwriting
|
||||
✓ creates skill directories with SKILL.md
|
||||
```
|
||||
|
||||
**File:** `tests/sync-copilot.test.ts`
|
||||
|
||||
Test cases (following `tests/sync-cursor.test.ts` pattern):
|
||||
|
||||
```
|
||||
describe("syncToCopilot")
|
||||
✓ symlinks skills to .github/skills/
|
||||
✓ skips skills with invalid names
|
||||
✓ merges MCP config with existing file
|
||||
✓ transforms MCP env var names to COPILOT_MCP_ prefix
|
||||
✓ writes MCP config with restricted permissions (0o600)
|
||||
```
|
||||
|
||||
### Phase 7: Documentation
|
||||
|
||||
**File:** `docs/specs/copilot.md`
|
||||
|
||||
Follow `docs/specs/cursor.md` format:
|
||||
- Last verified date
|
||||
- Primary sources (GitHub Docs URLs)
|
||||
- Config locations table
|
||||
- Agents section (`.agent.md` format, frontmatter fields)
|
||||
- Skills section (`SKILL.md` format)
|
||||
- MCP section (config structure, env var prefix requirement)
|
||||
- Character limits (30k agent body)
|
||||
|
||||
**File:** `README.md`
|
||||
|
||||
- Add "copilot" to the list of supported targets
|
||||
- Add usage example: `compound convert --to copilot ./plugins/compound-engineering`
|
||||
- Add sync example: `compound sync copilot`
|
||||
|
||||
## Acceptance Criteria
|
||||
|
||||
### Converter
|
||||
- [x] Agents convert to `.agent.md` with `description`, `tools: ["*"]`, `infer: true`
|
||||
- [x] Agent `model` passes through when present
|
||||
- [x] Agent `capabilities` fold into body as `## Capabilities`
|
||||
- [x] Missing description generates fallback
|
||||
- [x] Empty body generates fallback
|
||||
- [x] Body exceeding 30k chars triggers stderr warning
|
||||
- [x] Commands convert to SKILL.md format
|
||||
- [x] Command names flatten (`workflows:plan` → `plan`)
|
||||
- [x] Name collisions deduplicated with `-2`, `-3` suffix
|
||||
- [x] Command `allowed-tools` dropped silently
|
||||
- [x] Skills pass through as `CopilotSkillDir`
|
||||
- [x] MCP env vars prefixed with `COPILOT_MCP_`
|
||||
- [x] Already-prefixed env vars not double-prefixed
|
||||
- [x] MCP servers get `type` field (`local` or `sse`)
|
||||
- [x] Hooks trigger warning, skip conversion
|
||||
- [x] Content transformation: Task calls, slash commands, paths, @agent refs
|
||||
|
||||
### Writer
|
||||
- [x] Agents written to `.github/agents/{name}.agent.md`
|
||||
- [x] Generated skills written to `.github/skills/{name}/SKILL.md`
|
||||
- [x] Skill dirs copied to `.github/skills/{name}/`
|
||||
- [x] MCP config written to `.github/copilot-mcp-config.json`
|
||||
- [x] Existing MCP config backed up before overwrite
|
||||
- [x] No double-nesting when outputRoot is `.github`
|
||||
- [x] Empty bundles handled gracefully
|
||||
|
||||
### CLI Integration
|
||||
- [x] `compound convert --to copilot` works
|
||||
- [x] `compound sync copilot` works
|
||||
- [x] Copilot registered in `src/targets/index.ts`
|
||||
- [x] Sync resolves output to `.github/` in current directory
|
||||
|
||||
### Tests
|
||||
- [x] `tests/copilot-converter.test.ts` — all converter tests pass
|
||||
- [x] `tests/copilot-writer.test.ts` — all writer tests pass
|
||||
- [x] `tests/sync-copilot.test.ts` — all sync tests pass
|
||||
|
||||
### Documentation
|
||||
- [x] `docs/specs/copilot.md` — format specification
|
||||
- [x] `README.md` — updated with copilot target
|
||||
|
||||
## Files to Create
|
||||
|
||||
| File | Purpose |
|
||||
|------|---------|
|
||||
| `src/types/copilot.ts` | Type definitions |
|
||||
| `src/converters/claude-to-copilot.ts` | Converter logic |
|
||||
| `src/targets/copilot.ts` | Writer logic |
|
||||
| `src/sync/copilot.ts` | Sync handler |
|
||||
| `tests/copilot-converter.test.ts` | Converter tests |
|
||||
| `tests/copilot-writer.test.ts` | Writer tests |
|
||||
| `tests/sync-copilot.test.ts` | Sync tests |
|
||||
| `docs/specs/copilot.md` | Format specification |
|
||||
|
||||
## Files to Modify
|
||||
|
||||
| File | Change |
|
||||
|------|--------|
|
||||
| `src/targets/index.ts` | Register copilot target |
|
||||
| `src/commands/sync.ts` | Add copilot to valid targets, output root, switch case |
|
||||
| `README.md` | Add copilot to supported targets |
|
||||
|
||||
## References
|
||||
|
||||
- [Custom agents configuration - GitHub Docs](https://docs.github.com/en/copilot/reference/custom-agents-configuration)
|
||||
- [About Agent Skills - GitHub Docs](https://docs.github.com/en/copilot/concepts/agents/about-agent-skills)
|
||||
- [MCP and coding agent - GitHub Docs](https://docs.github.com/en/copilot/concepts/agents/coding-agent/mcp-and-coding-agent)
|
||||
- Existing converter: `src/converters/claude-to-cursor.ts`
|
||||
- Existing writer: `src/targets/cursor.ts`
|
||||
- Existing sync: `src/sync/cursor.ts`
|
||||
- Existing tests: `tests/cursor-converter.test.ts`, `tests/cursor-writer.test.ts`
|
||||
@@ -0,0 +1,360 @@
|
||||
---
|
||||
title: Auto-detect install targets and add Gemini sync
|
||||
type: feat
|
||||
status: completed
|
||||
date: 2026-02-14
|
||||
completed_date: 2026-02-14
|
||||
completed_by: "Claude Opus 4.6"
|
||||
actual_effort: "Completed in one session"
|
||||
---
|
||||
|
||||
# Auto-detect Install Targets and Add Gemini Sync
|
||||
|
||||
## Overview
|
||||
|
||||
Two related improvements to the converter CLI:
|
||||
|
||||
1. **`install --to all`** — Auto-detect which AI coding tools are installed and convert to all of them in one command
|
||||
2. **`sync --target gemini`** — Add Gemini CLI as a sync target (currently missing), then add `sync --target all` to sync personal config to every detected tool
|
||||
|
||||
## Problem Statement
|
||||
|
||||
Users currently must run 6 separate commands to install to all targets:
|
||||
|
||||
```bash
|
||||
bunx @every-env/compound-plugin install compound-engineering --to opencode
|
||||
bunx @every-env/compound-plugin install compound-engineering --to codex
|
||||
bunx @every-env/compound-plugin install compound-engineering --to droid
|
||||
bunx @every-env/compound-plugin install compound-engineering --to cursor
|
||||
bunx @every-env/compound-plugin install compound-engineering --to pi
|
||||
bunx @every-env/compound-plugin install compound-engineering --to gemini
|
||||
```
|
||||
|
||||
Similarly, sync requires separate commands per target. And Gemini sync doesn't exist yet.
|
||||
|
||||
## Acceptance Criteria
|
||||
|
||||
### Auto-detect install
|
||||
|
||||
- [x]`install --to all` detects installed tools and installs to each
|
||||
- [x]Detection checks config directories and/or binaries for each tool
|
||||
- [x]Prints which tools were detected and which were skipped
|
||||
- [x]Tools with no detection signal are skipped (not errored)
|
||||
- [x]`convert --to all` also works (same detection logic)
|
||||
- [x]Existing `--to <target>` behavior unchanged
|
||||
- [x]Tests for detection logic and `all` target handling
|
||||
|
||||
### Gemini sync
|
||||
|
||||
- [x]`sync --target gemini` symlinks skills and writes MCP servers to `.gemini/settings.json`
|
||||
- [x]MCP servers merged into existing `settings.json` (same pattern as writer)
|
||||
- [x]`gemini` added to `validTargets` in `sync.ts`
|
||||
- [x]Tests for Gemini sync
|
||||
|
||||
### Sync all
|
||||
|
||||
- [x]`sync --target all` syncs to all detected tools
|
||||
- [x]Reuses same detection logic as install
|
||||
- [x]Prints summary of what was synced where
|
||||
|
||||
## Implementation
|
||||
|
||||
### Phase 1: Tool Detection Utility
|
||||
|
||||
**Create `src/utils/detect-tools.ts`**
|
||||
|
||||
```typescript
|
||||
import os from "os"
|
||||
import path from "path"
|
||||
import { pathExists } from "./files"
|
||||
|
||||
export type DetectedTool = {
|
||||
name: string
|
||||
detected: boolean
|
||||
reason: string // e.g. "found ~/.codex/" or "not found"
|
||||
}
|
||||
|
||||
export async function detectInstalledTools(): Promise<DetectedTool[]> {
|
||||
const home = os.homedir()
|
||||
const cwd = process.cwd()
|
||||
|
||||
const checks: Array<{ name: string; paths: string[] }> = [
|
||||
{ name: "opencode", paths: [path.join(home, ".config", "opencode"), path.join(cwd, ".opencode")] },
|
||||
{ name: "codex", paths: [path.join(home, ".codex")] },
|
||||
{ name: "droid", paths: [path.join(home, ".factory")] },
|
||||
{ name: "cursor", paths: [path.join(cwd, ".cursor"), path.join(home, ".cursor")] },
|
||||
{ name: "pi", paths: [path.join(home, ".pi")] },
|
||||
{ name: "gemini", paths: [path.join(cwd, ".gemini"), path.join(home, ".gemini")] },
|
||||
]
|
||||
|
||||
const results: DetectedTool[] = []
|
||||
for (const check of checks) {
|
||||
let detected = false
|
||||
let reason = "not found"
|
||||
for (const p of check.paths) {
|
||||
if (await pathExists(p)) {
|
||||
detected = true
|
||||
reason = `found ${p}`
|
||||
break
|
||||
}
|
||||
}
|
||||
results.push({ name: check.name, detected, reason })
|
||||
}
|
||||
return results
|
||||
}
|
||||
|
||||
export async function getDetectedTargetNames(): Promise<string[]> {
|
||||
const tools = await detectInstalledTools()
|
||||
return tools.filter((t) => t.detected).map((t) => t.name)
|
||||
}
|
||||
```
|
||||
|
||||
**Detection heuristics:**
|
||||
|
||||
| Tool | Check paths | Notes |
|
||||
|------|------------|-------|
|
||||
| OpenCode | `~/.config/opencode/`, `.opencode/` | XDG config or project-local |
|
||||
| Codex | `~/.codex/` | Global only |
|
||||
| Droid | `~/.factory/` | Global only |
|
||||
| Cursor | `.cursor/`, `~/.cursor/` | Project-local or global |
|
||||
| Pi | `~/.pi/` | Global only |
|
||||
| Gemini | `.gemini/`, `~/.gemini/` | Project-local or global |
|
||||
|
||||
### Phase 2: Gemini Sync
|
||||
|
||||
**Create `src/sync/gemini.ts`**
|
||||
|
||||
Follow the Cursor sync pattern (`src/sync/cursor.ts`) since both use JSON config with `mcpServers` key:
|
||||
|
||||
```typescript
|
||||
import path from "path"
|
||||
import { symlinkSkills } from "../utils/symlink"
|
||||
import { backupFile, pathExists, readJson, writeJson } from "../utils/files"
|
||||
import type { ClaudeMcpServer } from "../types/claude"
|
||||
|
||||
export async function syncToGemini(
|
||||
skills: { name: string; sourceDir: string }[],
|
||||
mcpServers: Record<string, ClaudeMcpServer>,
|
||||
outputRoot: string,
|
||||
): Promise<void> {
|
||||
const geminiDir = path.join(outputRoot, ".gemini")
|
||||
|
||||
// Symlink skills
|
||||
if (skills.length > 0) {
|
||||
const skillsDir = path.join(geminiDir, "skills")
|
||||
await symlinkSkills(skills, skillsDir)
|
||||
}
|
||||
|
||||
// Merge MCP servers into settings.json
|
||||
if (Object.keys(mcpServers).length > 0) {
|
||||
const settingsPath = path.join(geminiDir, "settings.json")
|
||||
let existing: Record<string, unknown> = {}
|
||||
if (await pathExists(settingsPath)) {
|
||||
await backupFile(settingsPath)
|
||||
try {
|
||||
existing = await readJson<Record<string, unknown>>(settingsPath)
|
||||
} catch {
|
||||
console.warn("Warning: existing settings.json could not be parsed and will be replaced.")
|
||||
}
|
||||
}
|
||||
|
||||
const existingMcp = (existing.mcpServers && typeof existing.mcpServers === "object")
|
||||
? existing.mcpServers as Record<string, unknown>
|
||||
: {}
|
||||
|
||||
const merged = { ...existing, mcpServers: { ...existingMcp, ...convertMcpServers(mcpServers) } }
|
||||
await writeJson(settingsPath, merged)
|
||||
}
|
||||
}
|
||||
|
||||
function convertMcpServers(servers: Record<string, ClaudeMcpServer>) {
|
||||
const result: Record<string, Record<string, unknown>> = {}
|
||||
for (const [name, server] of Object.entries(servers)) {
|
||||
const entry: Record<string, unknown> = {}
|
||||
if (server.command) {
|
||||
entry.command = server.command
|
||||
if (server.args?.length) entry.args = server.args
|
||||
if (server.env && Object.keys(server.env).length > 0) entry.env = server.env
|
||||
} else if (server.url) {
|
||||
entry.url = server.url
|
||||
if (server.headers && Object.keys(server.headers).length > 0) entry.headers = server.headers
|
||||
}
|
||||
result[name] = entry
|
||||
}
|
||||
return result
|
||||
}
|
||||
```
|
||||
|
||||
**Update `src/commands/sync.ts`:**
|
||||
|
||||
- Add `"gemini"` to `validTargets` array
|
||||
- Import `syncToGemini` from `../sync/gemini`
|
||||
- Add case in switch for `"gemini"` calling `syncToGemini(skills, mcpServers, outputRoot)`
|
||||
|
||||
### Phase 3: Wire `--to all` into Install and Convert
|
||||
|
||||
**Modify `src/commands/install.ts`:**
|
||||
|
||||
```typescript
|
||||
import { detectInstalledTools } from "../utils/detect-tools"
|
||||
|
||||
// In args definition, update --to description:
|
||||
to: {
|
||||
type: "string",
|
||||
default: "opencode",
|
||||
description: "Target format (opencode | codex | droid | cursor | pi | gemini | all)",
|
||||
},
|
||||
|
||||
// In run(), before the existing target lookup:
|
||||
if (targetName === "all") {
|
||||
const detected = await detectInstalledTools()
|
||||
const activeTargets = detected.filter((t) => t.detected)
|
||||
|
||||
if (activeTargets.length === 0) {
|
||||
console.log("No AI coding tools detected. Install at least one tool first.")
|
||||
return
|
||||
}
|
||||
|
||||
console.log(`Detected ${activeTargets.length} tools:`)
|
||||
for (const tool of detected) {
|
||||
console.log(` ${tool.detected ? "✓" : "✗"} ${tool.name} — ${tool.reason}`)
|
||||
}
|
||||
|
||||
// Install to each detected target
|
||||
for (const tool of activeTargets) {
|
||||
const handler = targets[tool.name]
|
||||
const bundle = handler.convert(plugin, options)
|
||||
if (!bundle) continue
|
||||
const root = resolveTargetOutputRoot(tool.name, outputRoot, codexHome, piHome, hasExplicitOutput)
|
||||
await handler.write(root, bundle)
|
||||
console.log(`Installed ${plugin.manifest.name} to ${tool.name} at ${root}`)
|
||||
}
|
||||
|
||||
// Codex post-processing
|
||||
if (activeTargets.some((t) => t.name === "codex")) {
|
||||
await ensureCodexAgentsFile(codexHome)
|
||||
}
|
||||
return
|
||||
}
|
||||
```
|
||||
|
||||
**Same change in `src/commands/convert.ts`** with its version of `resolveTargetOutputRoot`.
|
||||
|
||||
### Phase 4: Wire `--target all` into Sync
|
||||
|
||||
**Modify `src/commands/sync.ts`:**
|
||||
|
||||
```typescript
|
||||
import { detectInstalledTools } from "../utils/detect-tools"
|
||||
|
||||
// Update validTargets:
|
||||
const validTargets = ["opencode", "codex", "pi", "droid", "cursor", "gemini", "all"] as const
|
||||
|
||||
// In run(), handle "all":
|
||||
if (targetName === "all") {
|
||||
const detected = await detectInstalledTools()
|
||||
const activeTargets = detected.filter((t) => t.detected).map((t) => t.name)
|
||||
|
||||
if (activeTargets.length === 0) {
|
||||
console.log("No AI coding tools detected.")
|
||||
return
|
||||
}
|
||||
|
||||
console.log(`Syncing to ${activeTargets.length} detected tools...`)
|
||||
for (const name of activeTargets) {
|
||||
// call existing sync logic for each target
|
||||
}
|
||||
return
|
||||
}
|
||||
```
|
||||
|
||||
### Phase 5: Tests
|
||||
|
||||
**Create `tests/detect-tools.test.ts`**
|
||||
|
||||
- Test detection with mocked directories (create temp dirs, check detection)
|
||||
- Test `getDetectedTargetNames` returns only detected tools
|
||||
- Test empty detection returns empty array
|
||||
|
||||
**Create `tests/gemini-sync.test.ts`**
|
||||
|
||||
Follow `tests/sync-cursor.test.ts` pattern:
|
||||
|
||||
- Test skills are symlinked to `.gemini/skills/`
|
||||
- Test MCP servers merged into `settings.json`
|
||||
- Test existing `settings.json` is backed up
|
||||
- Test empty skills/servers produce no output
|
||||
|
||||
**Update `tests/cli.test.ts`**
|
||||
|
||||
- Test `--to all` flag is accepted
|
||||
- Test `sync --target all` is accepted
|
||||
- Test `sync --target gemini` is accepted
|
||||
|
||||
### Phase 6: Documentation
|
||||
|
||||
**Update `README.md`:**
|
||||
|
||||
Add to install section:
|
||||
```bash
|
||||
# auto-detect installed tools and install to all
|
||||
bunx @every-env/compound-plugin install compound-engineering --to all
|
||||
```
|
||||
|
||||
Add to sync section:
|
||||
```bash
|
||||
# Sync to Gemini
|
||||
bunx @every-env/compound-plugin sync --target gemini
|
||||
|
||||
# Sync to all detected tools
|
||||
bunx @every-env/compound-plugin sync --target all
|
||||
```
|
||||
|
||||
## What We're NOT Doing
|
||||
|
||||
- Not adding binary detection (`which cursor`, `which gemini`) — directory checks are sufficient and don't require shell execution
|
||||
- Not adding interactive prompts ("Install to Cursor? y/n") — auto-detect is fire-and-forget
|
||||
- Not adding `--exclude` flag for skipping specific targets — can use `--to X --also Y` for manual selection
|
||||
- Not adding Gemini to the `sync` symlink watcher (no watcher exists for any target)
|
||||
|
||||
## Complexity Assessment
|
||||
|
||||
**Low-medium change.** All patterns are established:
|
||||
- Detection utility is new but simple (pathExists checks)
|
||||
- Gemini sync follows cursor sync pattern exactly
|
||||
- `--to all` is plumbing — iterate detected tools through existing handlers
|
||||
- No new dependencies needed
|
||||
|
||||
## References
|
||||
|
||||
- Cursor sync (reference pattern): `src/sync/cursor.ts`
|
||||
- Gemini writer (merge pattern): `src/targets/gemini.ts`
|
||||
- Install command: `src/commands/install.ts`
|
||||
- Sync command: `src/commands/sync.ts`
|
||||
- File utilities: `src/utils/files.ts`
|
||||
- Symlink utilities: `src/utils/symlink.ts`
|
||||
|
||||
## Completion Summary
|
||||
|
||||
### What Was Delivered
|
||||
- Tool detection utility (`src/utils/detect-tools.ts`) with `detectInstalledTools()` and `getDetectedTargetNames()`
|
||||
- Gemini sync (`src/sync/gemini.ts`) following cursor sync pattern — symlinks skills, merges MCP servers into `settings.json`
|
||||
- `install --to all` and `convert --to all` auto-detect and install to all detected tools
|
||||
- `sync --target gemini` added to sync command
|
||||
- `sync --target all` syncs to all detected tools with summary output
|
||||
- 8 new tests across 2 test files (detect-tools + sync-gemini)
|
||||
|
||||
### Implementation Statistics
|
||||
- 4 new files, 3 modified files
|
||||
- 139 tests passing (8 new + 131 existing)
|
||||
- No new dependencies
|
||||
|
||||
### Git Commits
|
||||
- `e4d730d` feat: add detect-tools utility and Gemini sync with tests
|
||||
- `bc655f7` feat: wire --to all into install/convert and --target all/gemini into sync
|
||||
- `877e265` docs: add auto-detect and Gemini sync to README, bump to 0.8.0
|
||||
|
||||
### Completion Details
|
||||
- **Completed By:** Claude Opus 4.6
|
||||
- **Date:** 2026-02-14
|
||||
- **Session:** Single session, TDD approach
|
||||
627
docs/plans/2026-02-25-feat-windsurf-global-scope-support-plan.md
Normal file
627
docs/plans/2026-02-25-feat-windsurf-global-scope-support-plan.md
Normal file
@@ -0,0 +1,627 @@
|
||||
---
|
||||
title: Windsurf Global Scope Support
|
||||
type: feat
|
||||
status: completed
|
||||
date: 2026-02-25
|
||||
deepened: 2026-02-25
|
||||
prior: docs/plans/2026-02-23-feat-add-windsurf-target-provider-plan.md (removed — superseded)
|
||||
---
|
||||
|
||||
# Windsurf Global Scope Support
|
||||
|
||||
## Post-Implementation Revisions (2026-02-26)
|
||||
|
||||
After auditing the implementation against `docs/specs/windsurf.md`, two significant changes were made:
|
||||
|
||||
1. **Agents → Skills (not Workflows)**: Claude agents map to Windsurf Skills (`skills/{name}/SKILL.md`), not Workflows. Skills are "complex multi-step tasks with supporting resources" — a better conceptual match for specialized expertise/personas. Workflows are "reusable step-by-step procedures" — a better match for Claude Commands (slash commands).
|
||||
|
||||
2. **Workflows are flat files**: Command workflows are written to `global_workflows/{name}.md` (global scope) or `workflows/{name}.md` (workspace scope). No subdirectories — the spec requires flat files.
|
||||
|
||||
3. **Content transforms updated**: `@agent-name` references are kept as-is (Windsurf skill invocation syntax). `/command` references produce `/{name}` (not `/commands/{name}`). `Task agent(args)` produces `Use the @agent-name skill: args`.
|
||||
|
||||
### Final Component Mapping (per spec)
|
||||
|
||||
| Claude Code | Windsurf | Output Path | Invocation |
|
||||
|---|---|---|---|
|
||||
| Agents (`.md`) | Skills | `skills/{name}/SKILL.md` | `@skill-name` or automatic |
|
||||
| Commands (`.md`) | Workflows (flat) | `global_workflows/{name}.md` (global) / `workflows/{name}.md` (workspace) | `/{workflow-name}` |
|
||||
| Skills (`SKILL.md`) | Skills (pass-through) | `skills/{name}/SKILL.md` | `@skill-name` |
|
||||
| MCP servers | `mcp_config.json` | `mcp_config.json` | N/A |
|
||||
| Hooks | Skipped with warning | N/A | N/A |
|
||||
| CLAUDE.md | Skipped | N/A | N/A |
|
||||
|
||||
### Files Changed in Revision
|
||||
|
||||
- `src/types/windsurf.ts` — `agentWorkflows` → `agentSkills: WindsurfGeneratedSkill[]`
|
||||
- `src/converters/claude-to-windsurf.ts` — `convertAgentToSkill()`, updated content transforms
|
||||
- `src/targets/windsurf.ts` — Skills written as `skills/{name}/SKILL.md`, flat workflows
|
||||
- Tests updated to match
|
||||
|
||||
---
|
||||
|
||||
## Enhancement Summary
|
||||
|
||||
**Deepened on:** 2026-02-25
|
||||
**Research agents used:** architecture-strategist, kieran-typescript-reviewer, security-sentinel, code-simplicity-reviewer, pattern-recognition-specialist
|
||||
**External research:** Windsurf MCP docs, Windsurf tutorial docs
|
||||
|
||||
### Key Improvements from Deepening
|
||||
1. **HTTP/SSE servers should be INCLUDED** — Windsurf supports all 3 transport types (stdio, Streamable HTTP, SSE). Original plan incorrectly skipped them.
|
||||
2. **File permissions: use `0o600`** — `mcp_config.json` contains secrets and must not be world-readable. Add secure write support.
|
||||
3. **Extract `resolveTargetOutputRoot` to shared utility** — both commands duplicate this; adding scope makes it worse. Extract first.
|
||||
4. **Bug fix: missing `result[name] = entry`** — all 5 review agents caught a copy-paste bug in the `buildMcpConfig` sample code.
|
||||
5. **`hasPotentialSecrets` to shared utility** — currently in sync.ts, would be duplicated. Extract to `src/utils/secrets.ts`.
|
||||
6. **Windsurf `mcp_config.json` is global-only** — per Windsurf docs, no per-project MCP config support. Workspace scope writes it for forward-compatibility but emit a warning.
|
||||
7. **Windsurf supports `${env:VAR}` interpolation** — consider writing env var references instead of literal values for secrets.
|
||||
|
||||
### New Considerations Discovered
|
||||
- Backup files accumulate with secrets and are never cleaned up — cap at 3 backups
|
||||
- Workspace `mcp_config.json` could be committed to git — warn about `.gitignore`
|
||||
- `WindsurfMcpServerEntry` type needs `serverUrl` field for HTTP/SSE servers
|
||||
- Simplicity reviewer recommends handling scope as windsurf-specific in CLI rather than generic `TargetHandler` fields — but brainstorm explicitly chose "generic with windsurf as first adopter". **Decision: keep generic approach** per user's brainstorm decision, with JSDoc documenting the relationship between `defaultScope` and `supportedScopes`.
|
||||
|
||||
---
|
||||
|
||||
## Overview
|
||||
|
||||
Add a generic `--scope global|workspace` flag to the converter CLI with Windsurf as the first adopter. Global scope writes to `~/.codeium/windsurf/`, making workflows, skills, and MCP servers available across all projects. This also upgrades MCP handling from a human-readable setup doc (`mcp-setup.md`) to a proper machine-readable config (`mcp_config.json`), and removes AGENTS.md generation (the plugin's CLAUDE.md contains development-internal instructions, not user-facing content).
|
||||
|
||||
## Problem Statement / Motivation
|
||||
|
||||
The current Windsurf converter (v0.10.0) writes everything to project-level `.windsurf/`, requiring re-installation per project. Windsurf supports global paths for skills (`~/.codeium/windsurf/skills/`) and MCP config (`~/.codeium/windsurf/mcp_config.json`). Users should install once and get capabilities everywhere.
|
||||
|
||||
Additionally, the v0.10.0 MCP output was a markdown setup guide — not an actual integration. Windsurf reads `mcp_config.json` directly, so we should write to that file.
|
||||
|
||||
## Breaking Changes from v0.10.0
|
||||
|
||||
This is a **minor version bump** (v0.11.0) with intentional breaking changes to the experimental Windsurf target:
|
||||
|
||||
1. **Default output location changed** — `--to windsurf` now defaults to global scope (`~/.codeium/windsurf/`). Use `--scope workspace` for the old behavior.
|
||||
2. **AGENTS.md no longer generated** — old files are left in place (not deleted).
|
||||
3. **`mcp-setup.md` replaced by `mcp_config.json`** — proper machine-readable integration. Old files left in place.
|
||||
4. **Env var secrets included with warning** — previously redacted, now included (required for the config file to work).
|
||||
5. **`--output` semantics changed** — `--output` now specifies the direct target directory (not a parent where `.windsurf/` is created).
|
||||
|
||||
## Proposed Solution
|
||||
|
||||
### Phase 0: Extract Shared Utilities (prerequisite)
|
||||
|
||||
**Files:** `src/utils/resolve-output.ts` (new), `src/utils/secrets.ts` (new)
|
||||
|
||||
#### 0a. Extract `resolveTargetOutputRoot` to shared utility
|
||||
|
||||
Both `install.ts` and `convert.ts` have near-identical `resolveTargetOutputRoot` functions that are already diverging (`hasExplicitOutput` exists in install.ts but not convert.ts). Adding scope would make the duplication worse.
|
||||
|
||||
- [x] Create `src/utils/resolve-output.ts` with a unified function:
|
||||
|
||||
```typescript
|
||||
import os from "os"
|
||||
import path from "path"
|
||||
import type { TargetScope } from "../targets"
|
||||
|
||||
export function resolveTargetOutputRoot(options: {
|
||||
targetName: string
|
||||
outputRoot: string
|
||||
codexHome: string
|
||||
piHome: string
|
||||
hasExplicitOutput: boolean
|
||||
scope?: TargetScope
|
||||
}): string {
|
||||
const { targetName, outputRoot, codexHome, piHome, hasExplicitOutput, scope } = options
|
||||
if (targetName === "codex") return codexHome
|
||||
if (targetName === "pi") return piHome
|
||||
if (targetName === "droid") return path.join(os.homedir(), ".factory")
|
||||
if (targetName === "cursor") {
|
||||
const base = hasExplicitOutput ? outputRoot : process.cwd()
|
||||
return path.join(base, ".cursor")
|
||||
}
|
||||
if (targetName === "gemini") {
|
||||
const base = hasExplicitOutput ? outputRoot : process.cwd()
|
||||
return path.join(base, ".gemini")
|
||||
}
|
||||
if (targetName === "copilot") {
|
||||
const base = hasExplicitOutput ? outputRoot : process.cwd()
|
||||
return path.join(base, ".github")
|
||||
}
|
||||
if (targetName === "kiro") {
|
||||
const base = hasExplicitOutput ? outputRoot : process.cwd()
|
||||
return path.join(base, ".kiro")
|
||||
}
|
||||
if (targetName === "windsurf") {
|
||||
if (hasExplicitOutput) return outputRoot
|
||||
if (scope === "global") return path.join(os.homedir(), ".codeium", "windsurf")
|
||||
return path.join(process.cwd(), ".windsurf")
|
||||
}
|
||||
return outputRoot
|
||||
}
|
||||
```
|
||||
|
||||
- [x] Update `install.ts` to import and call `resolveTargetOutputRoot` from shared utility
|
||||
- [x] Update `convert.ts` to import and call `resolveTargetOutputRoot` from shared utility
|
||||
- [x] Add `hasExplicitOutput` tracking to `convert.ts` (currently missing)
|
||||
|
||||
### Research Insights (Phase 0)
|
||||
|
||||
**Architecture review:** Both commands will call the same function with the same signature. This eliminates the divergence and ensures scope resolution has a single source of truth. The `--also` loop in both commands also uses this function with `handler.defaultScope`.
|
||||
|
||||
**Pattern review:** This follows the same extraction pattern as `resolveTargetHome` in `src/utils/resolve-home.ts`.
|
||||
|
||||
#### 0b. Extract `hasPotentialSecrets` to shared utility
|
||||
|
||||
Currently in `sync.ts:20-31`. The same regex pattern also appears in `claude-to-windsurf.ts:223` as `redactEnvValue`. Extract to avoid a third copy.
|
||||
|
||||
- [x] Create `src/utils/secrets.ts`:
|
||||
|
||||
```typescript
|
||||
const SENSITIVE_PATTERN = /key|token|secret|password|credential|api_key/i
|
||||
|
||||
export function hasPotentialSecrets(
|
||||
servers: Record<string, { env?: Record<string, string> }>,
|
||||
): boolean {
|
||||
for (const server of Object.values(servers)) {
|
||||
if (server.env) {
|
||||
for (const key of Object.keys(server.env)) {
|
||||
if (SENSITIVE_PATTERN.test(key)) return true
|
||||
}
|
||||
}
|
||||
}
|
||||
return false
|
||||
}
|
||||
```
|
||||
|
||||
- [x] Update `sync.ts` to import from shared utility
|
||||
- [x] Use in new windsurf converter
|
||||
|
||||
### Phase 1: Types and TargetHandler
|
||||
|
||||
**Files:** `src/types/windsurf.ts`, `src/targets/index.ts`
|
||||
|
||||
#### 1a. Update WindsurfBundle type
|
||||
|
||||
```typescript
|
||||
// src/types/windsurf.ts
|
||||
export type WindsurfMcpServerEntry = {
|
||||
command?: string
|
||||
args?: string[]
|
||||
env?: Record<string, string>
|
||||
serverUrl?: string
|
||||
headers?: Record<string, string>
|
||||
}
|
||||
|
||||
export type WindsurfMcpConfig = {
|
||||
mcpServers: Record<string, WindsurfMcpServerEntry>
|
||||
}
|
||||
|
||||
export type WindsurfBundle = {
|
||||
agentWorkflows: WindsurfWorkflow[]
|
||||
commandWorkflows: WindsurfWorkflow[]
|
||||
skillDirs: WindsurfSkillDir[]
|
||||
mcpConfig: WindsurfMcpConfig | null
|
||||
}
|
||||
```
|
||||
|
||||
- [x] Remove `agentsMd: string | null`
|
||||
- [x] Replace `mcpSetupDoc: string | null` with `mcpConfig: WindsurfMcpConfig | null`
|
||||
- [x] Add `WindsurfMcpServerEntry` (supports both stdio and HTTP/SSE) and `WindsurfMcpConfig` types
|
||||
|
||||
### Research Insights (Phase 1a)
|
||||
|
||||
**Windsurf docs confirm** three transport types: stdio (`command` + `args`), Streamable HTTP (`serverUrl`), and SSE (`serverUrl` or `url`). The `WindsurfMcpServerEntry` type must support all three — making `command` optional and adding `serverUrl` and `headers` fields.
|
||||
|
||||
**TypeScript reviewer:** Consider making `WindsurfMcpServerEntry` a discriminated union if strict typing is desired. However, since this mirrors JSON config structure, a flat type with optional fields is pragmatically simpler.
|
||||
|
||||
#### 1b. Add TargetScope to TargetHandler
|
||||
|
||||
```typescript
|
||||
// src/targets/index.ts
|
||||
export type TargetScope = "global" | "workspace"
|
||||
|
||||
export type TargetHandler<TBundle = unknown> = {
|
||||
name: string
|
||||
implemented: boolean
|
||||
/**
|
||||
* Default scope when --scope is not provided.
|
||||
* Only meaningful when supportedScopes is defined.
|
||||
* Falls back to "workspace" if absent.
|
||||
*/
|
||||
defaultScope?: TargetScope
|
||||
/** Valid scope values. If absent, the --scope flag is rejected for this target. */
|
||||
supportedScopes?: TargetScope[]
|
||||
convert: (plugin: ClaudePlugin, options: ClaudeToOpenCodeOptions) => TBundle | null
|
||||
write: (outputRoot: string, bundle: TBundle) => Promise<void>
|
||||
}
|
||||
```
|
||||
|
||||
- [x] Add `TargetScope` type export
|
||||
- [x] Add `defaultScope?` and `supportedScopes?` to `TargetHandler` with JSDoc
|
||||
- [x] Set windsurf target: `defaultScope: "global"`, `supportedScopes: ["global", "workspace"]`
|
||||
- [x] No changes to other targets (they have no scope fields, flag is ignored)
|
||||
|
||||
### Research Insights (Phase 1b)
|
||||
|
||||
**Simplicity review:** Argued this is premature generalization (only 1 of 8 targets uses scopes). Recommended handling scope as windsurf-specific with `if (targetName !== "windsurf")` guard instead. **Decision: keep generic approach** per brainstorm decision "Generic with windsurf as first adopter", but add JSDoc documenting the invariant.
|
||||
|
||||
**TypeScript review:** Suggested a `ScopeConfig` grouped object to prevent `defaultScope` without `supportedScopes`. The JSDoc approach is simpler and sufficient for now.
|
||||
|
||||
**Architecture review:** Adding optional fields to `TargetHandler` follows Open/Closed Principle — existing targets are unaffected. Clean extension.
|
||||
|
||||
### Phase 2: Converter Changes
|
||||
|
||||
**Files:** `src/converters/claude-to-windsurf.ts`
|
||||
|
||||
#### 2a. Remove AGENTS.md generation
|
||||
|
||||
- [x] Remove `buildAgentsMd()` function
|
||||
- [x] Remove `agentsMd` from return value
|
||||
|
||||
#### 2b. Replace MCP setup doc with MCP config
|
||||
|
||||
- [x] Remove `buildMcpSetupDoc()` function
|
||||
- [x] Remove `redactEnvValue()` helper
|
||||
- [x] Add `buildMcpConfig()` that returns `WindsurfMcpConfig | null`
|
||||
- [x] Include **all** env vars (including secrets) — no redaction
|
||||
- [x] Use shared `hasPotentialSecrets()` from `src/utils/secrets.ts`
|
||||
- [x] Include **both** stdio and HTTP/SSE servers (Windsurf supports all transport types)
|
||||
|
||||
```typescript
|
||||
function buildMcpConfig(
|
||||
servers?: Record<string, ClaudeMcpServer>,
|
||||
): WindsurfMcpConfig | null {
|
||||
if (!servers || Object.keys(servers).length === 0) return null
|
||||
|
||||
const result: Record<string, WindsurfMcpServerEntry> = {}
|
||||
for (const [name, server] of Object.entries(servers)) {
|
||||
if (server.command) {
|
||||
// stdio transport
|
||||
const entry: WindsurfMcpServerEntry = { command: server.command }
|
||||
if (server.args?.length) entry.args = server.args
|
||||
if (server.env && Object.keys(server.env).length > 0) entry.env = server.env
|
||||
result[name] = entry
|
||||
} else if (server.url) {
|
||||
// HTTP/SSE transport
|
||||
const entry: WindsurfMcpServerEntry = { serverUrl: server.url }
|
||||
if (server.headers && Object.keys(server.headers).length > 0) entry.headers = server.headers
|
||||
if (server.env && Object.keys(server.env).length > 0) entry.env = server.env
|
||||
result[name] = entry
|
||||
} else {
|
||||
console.warn(`Warning: MCP server "${name}" has no command or URL. Skipping.`)
|
||||
continue
|
||||
}
|
||||
}
|
||||
|
||||
if (Object.keys(result).length === 0) return null
|
||||
|
||||
// Warn about secrets (don't redact — they're needed for the config to work)
|
||||
if (hasPotentialSecrets(result)) {
|
||||
console.warn(
|
||||
"Warning: MCP servers contain env vars that may include secrets (API keys, tokens).\n" +
|
||||
" These will be written to mcp_config.json. Review before sharing the config file.",
|
||||
)
|
||||
}
|
||||
|
||||
return { mcpServers: result }
|
||||
}
|
||||
```
|
||||
|
||||
### Research Insights (Phase 2)
|
||||
|
||||
**Windsurf docs (critical correction):** Windsurf supports **stdio, Streamable HTTP, and SSE** transports in `mcp_config.json`. HTTP/SSE servers use `serverUrl` (not `url`). The original plan incorrectly planned to skip HTTP/SSE servers. This is now corrected — all transport types are included.
|
||||
|
||||
**All 5 review agents flagged:** The original code sample was missing `result[name] = entry` — the entry was built but never stored. Fixed above.
|
||||
|
||||
**Security review:** The warning message should enumerate which specific env var names triggered detection. Enhanced version:
|
||||
|
||||
```typescript
|
||||
if (hasPotentialSecrets(result)) {
|
||||
const flagged = Object.entries(result)
|
||||
.filter(([, s]) => s.env && Object.keys(s.env).some(k => SENSITIVE_PATTERN.test(k)))
|
||||
.map(([name]) => name)
|
||||
console.warn(
|
||||
`Warning: MCP servers contain env vars that may include secrets: ${flagged.join(", ")}.\n` +
|
||||
" These will be written to mcp_config.json. Review before sharing the config file.",
|
||||
)
|
||||
}
|
||||
```
|
||||
|
||||
**Windsurf env var interpolation:** Windsurf supports `${env:VARIABLE_NAME}` syntax in `mcp_config.json`. Future enhancement: write env var references instead of literal values for secrets. Out of scope for v0.11.0 (requires more research on which fields support interpolation).
|
||||
|
||||
### Phase 3: Writer Changes
|
||||
|
||||
**Files:** `src/targets/windsurf.ts`, `src/utils/files.ts`
|
||||
|
||||
#### 3a. Simplify writer — remove AGENTS.md and double-nesting guard
|
||||
|
||||
The writer always writes directly into `outputRoot`. The CLI resolves the correct output root based on scope.
|
||||
|
||||
- [x] Remove AGENTS.md writing block (lines 10-17)
|
||||
- [x] Remove `resolveWindsurfPaths()` — no longer needed
|
||||
- [x] Write workflows, skills, and MCP config directly into `outputRoot`
|
||||
|
||||
### Research Insights (Phase 3a)
|
||||
|
||||
**Pattern review (dissent):** Every other writer (kiro, copilot, gemini, droid) has a `resolve*Paths()` function with a double-nesting guard. Removing it makes Windsurf the only target where the CLI fully owns nesting. This creates an inconsistency in the `write()` contract.
|
||||
|
||||
**Resolution:** Accept the divergence — Windsurf has genuinely different semantics (global vs workspace). Add a JSDoc comment on `TargetHandler.write()` documenting that some writers may apply additional nesting while the Windsurf writer expects the final resolved path. Long-term, other targets could migrate to this pattern in a separate refactor.
|
||||
|
||||
#### 3b. Replace MCP setup doc with JSON config merge
|
||||
|
||||
Follow Kiro pattern (`src/targets/kiro.ts:68-92`) with security hardening:
|
||||
|
||||
- [x] Read existing `mcp_config.json` if present
|
||||
- [x] Backup before overwrite (`backupFile()`)
|
||||
- [x] Parse existing JSON (warn and replace if corrupted; add `!Array.isArray()` guard)
|
||||
- [x] Merge at `mcpServers` key: plugin entries overwrite same-name entries, user entries preserved
|
||||
- [x] Preserve all other top-level keys in existing file
|
||||
- [x] Write merged result with **restrictive permissions** (`0o600`)
|
||||
- [x] Emit warning when writing to workspace scope (Windsurf `mcp_config.json` is global-only per docs)
|
||||
|
||||
```typescript
|
||||
// MCP config merge with security hardening
|
||||
if (bundle.mcpConfig) {
|
||||
const mcpPath = path.join(outputRoot, "mcp_config.json")
|
||||
const backupPath = await backupFile(mcpPath)
|
||||
if (backupPath) {
|
||||
console.log(`Backed up existing mcp_config.json to ${backupPath}`)
|
||||
}
|
||||
|
||||
let existingConfig: Record<string, unknown> = {}
|
||||
if (await pathExists(mcpPath)) {
|
||||
try {
|
||||
const parsed = await readJson<unknown>(mcpPath)
|
||||
if (parsed && typeof parsed === "object" && !Array.isArray(parsed)) {
|
||||
existingConfig = parsed as Record<string, unknown>
|
||||
}
|
||||
} catch {
|
||||
console.warn("Warning: existing mcp_config.json could not be parsed and will be replaced.")
|
||||
}
|
||||
}
|
||||
|
||||
const existingServers =
|
||||
existingConfig.mcpServers &&
|
||||
typeof existingConfig.mcpServers === "object" &&
|
||||
!Array.isArray(existingConfig.mcpServers)
|
||||
? (existingConfig.mcpServers as Record<string, unknown>)
|
||||
: {}
|
||||
const merged = { ...existingConfig, mcpServers: { ...existingServers, ...bundle.mcpConfig.mcpServers } }
|
||||
await writeJsonSecure(mcpPath, merged) // 0o600 permissions
|
||||
}
|
||||
```
|
||||
|
||||
### Research Insights (Phase 3b)
|
||||
|
||||
**Security review (HIGH):** The current `writeJson()` in `src/utils/files.ts` uses default umask (`0o644`) — world-readable. The sync targets all use `{ mode: 0o600 }` for secret-containing files. The Windsurf writer (and Kiro writer) must do the same.
|
||||
|
||||
**Implementation:** Add a `writeJsonSecure()` helper or add a `mode` parameter to `writeJson()`:
|
||||
|
||||
```typescript
|
||||
// src/utils/files.ts
|
||||
export async function writeJsonSecure(filePath: string, data: unknown): Promise<void> {
|
||||
const content = JSON.stringify(data, null, 2)
|
||||
await ensureDir(path.dirname(filePath))
|
||||
await fs.writeFile(filePath, content + "\n", { encoding: "utf8", mode: 0o600 })
|
||||
}
|
||||
```
|
||||
|
||||
**Security review (MEDIUM):** Backup files inherit default permissions. Ensure `backupFile()` also sets `0o600` on the backup copy when the source may contain secrets.
|
||||
|
||||
**Security review (MEDIUM):** Workspace `mcp_config.json` could be committed to git. After writing to workspace scope, emit a warning:
|
||||
|
||||
```
|
||||
Warning: .windsurf/mcp_config.json may contain secrets. Ensure it is in .gitignore.
|
||||
```
|
||||
|
||||
**TypeScript review:** The `readJson<Record<string, unknown>>` assertion is unsafe — a valid JSON array or string passes parsing but fails the type. Added `!Array.isArray()` guard.
|
||||
|
||||
**TypeScript review:** The `bundle.mcpConfig` null check is sufficient — when non-null, `mcpServers` is guaranteed to have entries (the converter returns null for empty servers). Simplified from `bundle.mcpConfig && Object.keys(...)`.
|
||||
|
||||
**Windsurf docs (important):** `mcp_config.json` is a **global configuration only** — Windsurf has no per-project MCP config support. Writing it to `.windsurf/` in workspace scope may not be discovered by Windsurf. Emit a warning for workspace scope but still write the file for forward-compatibility.
|
||||
|
||||
#### 3c. Updated writer structure
|
||||
|
||||
```typescript
|
||||
export async function writeWindsurfBundle(outputRoot: string, bundle: WindsurfBundle): Promise<void> {
|
||||
await ensureDir(outputRoot)
|
||||
|
||||
// Write agent workflows
|
||||
if (bundle.agentWorkflows.length > 0) {
|
||||
const agentDir = path.join(outputRoot, "workflows", "agents")
|
||||
await ensureDir(agentDir)
|
||||
for (const workflow of bundle.agentWorkflows) {
|
||||
validatePathSafe(workflow.name, "agent workflow")
|
||||
const content = formatFrontmatter({ description: workflow.description }, `# ${workflow.name}\n\n${workflow.body}`)
|
||||
await writeText(path.join(agentDir, `${workflow.name}.md`), content + "\n")
|
||||
}
|
||||
}
|
||||
|
||||
// Write command workflows
|
||||
if (bundle.commandWorkflows.length > 0) {
|
||||
const cmdDir = path.join(outputRoot, "workflows", "commands")
|
||||
await ensureDir(cmdDir)
|
||||
for (const workflow of bundle.commandWorkflows) {
|
||||
validatePathSafe(workflow.name, "command workflow")
|
||||
const content = formatFrontmatter({ description: workflow.description }, `# ${workflow.name}\n\n${workflow.body}`)
|
||||
await writeText(path.join(cmdDir, `${workflow.name}.md`), content + "\n")
|
||||
}
|
||||
}
|
||||
|
||||
// Copy skill directories
|
||||
if (bundle.skillDirs.length > 0) {
|
||||
const skillsDir = path.join(outputRoot, "skills")
|
||||
await ensureDir(skillsDir)
|
||||
for (const skill of bundle.skillDirs) {
|
||||
validatePathSafe(skill.name, "skill directory")
|
||||
const destDir = path.join(skillsDir, skill.name)
|
||||
const resolvedDest = path.resolve(destDir)
|
||||
if (!resolvedDest.startsWith(path.resolve(skillsDir))) {
|
||||
console.warn(`Warning: Skill name "${skill.name}" escapes skills/. Skipping.`)
|
||||
continue
|
||||
}
|
||||
await copyDir(skill.sourceDir, destDir)
|
||||
}
|
||||
}
|
||||
|
||||
// Merge MCP config (see 3b above)
|
||||
if (bundle.mcpConfig) {
|
||||
// ... merge logic from 3b
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
### Phase 4: CLI Wiring
|
||||
|
||||
**Files:** `src/commands/install.ts`, `src/commands/convert.ts`
|
||||
|
||||
#### 4a. Add `--scope` flag to both commands
|
||||
|
||||
```typescript
|
||||
scope: {
|
||||
type: "string",
|
||||
description: "Scope level: global | workspace (default varies by target)",
|
||||
},
|
||||
```
|
||||
|
||||
- [x] Add `scope` arg to `install.ts`
|
||||
- [x] Add `scope` arg to `convert.ts`
|
||||
|
||||
#### 4b. Validate scope with type guard
|
||||
|
||||
Use a proper type guard instead of unsafe `as TargetScope` cast:
|
||||
|
||||
```typescript
|
||||
function isTargetScope(value: string): value is TargetScope {
|
||||
return value === "global" || value === "workspace"
|
||||
}
|
||||
|
||||
const scopeValue = args.scope ? String(args.scope) : undefined
|
||||
if (scopeValue !== undefined) {
|
||||
if (!target.supportedScopes) {
|
||||
throw new Error(`Target "${targetName}" does not support the --scope flag.`)
|
||||
}
|
||||
if (!isTargetScope(scopeValue) || !target.supportedScopes.includes(scopeValue)) {
|
||||
throw new Error(`Target "${targetName}" does not support --scope ${scopeValue}. Supported: ${target.supportedScopes.join(", ")}`)
|
||||
}
|
||||
}
|
||||
const resolvedScope = scopeValue ?? target.defaultScope ?? "workspace"
|
||||
```
|
||||
|
||||
- [x] Add `isTargetScope` type guard
|
||||
- [x] Add scope validation in both commands (single block, not two separate checks)
|
||||
|
||||
### Research Insights (Phase 4b)
|
||||
|
||||
**TypeScript review:** The original plan cast `scopeValue as TargetScope` before validation — a type lie. Use a proper type guard function to keep the type system honest.
|
||||
|
||||
**Simplicity review:** The two-step validation (check supported, then check exists) can be a single block with the type guard approach above.
|
||||
|
||||
#### 4c. Update output root resolution
|
||||
|
||||
Both commands now use the shared `resolveTargetOutputRoot` from Phase 0a.
|
||||
|
||||
- [x] Call shared function with `scope: resolvedScope` for primary target
|
||||
- [x] Default scope: `target.defaultScope ?? "workspace"` (only used when target supports scopes)
|
||||
|
||||
#### 4d. Handle `--also` targets
|
||||
|
||||
`--scope` applies only to the primary `--to` target. Extra `--also` targets use their own `defaultScope`.
|
||||
|
||||
- [x] Pass `handler.defaultScope` for `--also` targets (each uses its own default)
|
||||
- [x] Update the `--also` loop in both commands to use target-specific scope resolution
|
||||
|
||||
### Research Insights (Phase 4d)
|
||||
|
||||
**Architecture review:** There is no way for users to specify scope for an `--also` target (e.g., `--also windsurf:workspace`). Accept as a known v0.11.0 limitation. If users need workspace scope for windsurf, they can run two separate commands. Add a code comment indicating where per-target scope overrides would be added in the future.
|
||||
|
||||
### Phase 5: Tests
|
||||
|
||||
**Files:** `tests/windsurf-converter.test.ts`, `tests/windsurf-writer.test.ts`
|
||||
|
||||
#### 5a. Update converter tests
|
||||
|
||||
- [x] Remove all AGENTS.md tests (lines 275-303: empty plugin, CLAUDE.md missing)
|
||||
- [x] Remove all `mcpSetupDoc` tests (lines 305-366: stdio, HTTP/SSE, redaction, null)
|
||||
- [x] Update `fixturePlugin` default — remove `agentsMd` and `mcpSetupDoc` references
|
||||
- [x] Add `mcpConfig` tests:
|
||||
- stdio server produces correct JSON structure with `command`, `args`, `env`
|
||||
- HTTP/SSE server produces correct JSON structure with `serverUrl`, `headers`
|
||||
- mixed servers (stdio + HTTP) both included
|
||||
- env vars included (not redacted) — verify actual values present
|
||||
- `hasPotentialSecrets()` emits console.warn for sensitive keys
|
||||
- `hasPotentialSecrets()` does NOT warn when no sensitive keys
|
||||
- no servers produces null mcpConfig
|
||||
- empty bundle has null mcpConfig
|
||||
- server with no command and no URL is skipped with warning
|
||||
|
||||
#### 5b. Update writer tests
|
||||
|
||||
- [x] Remove AGENTS.md tests (backup test, creation test, double-nesting AGENTS.md parent test)
|
||||
- [x] Remove double-nesting guard test (guard removed)
|
||||
- [x] Remove `mcp-setup.md` write test
|
||||
- [x] Update `emptyBundle` fixture — remove `agentsMd`, `mcpSetupDoc`, add `mcpConfig: null`
|
||||
- [x] Add `mcp_config.json` tests:
|
||||
- writes mcp_config.json to outputRoot
|
||||
- merges with existing mcp_config.json (preserves user servers)
|
||||
- backs up existing mcp_config.json before overwrite
|
||||
- handles corrupted existing mcp_config.json (warn and replace)
|
||||
- handles existing mcp_config.json with array (not object) at root
|
||||
- handles existing mcp_config.json with `mcpServers: null`
|
||||
- preserves non-mcpServers keys in existing file
|
||||
- server name collision: plugin entry wins
|
||||
- file permissions are 0o600 (not world-readable)
|
||||
- [x] Update full bundle test — writer writes directly into outputRoot (no `.windsurf/` nesting)
|
||||
|
||||
#### 5c. Add scope resolution tests
|
||||
|
||||
Test the shared `resolveTargetOutputRoot` function:
|
||||
|
||||
- [x] Default scope for windsurf is "global" → resolves to `~/.codeium/windsurf/`
|
||||
- [x] Explicit `--scope workspace` → resolves to `cwd/.windsurf/`
|
||||
- [x] `--output` overrides scope resolution (both global and workspace)
|
||||
- [x] Invalid scope value for windsurf → error
|
||||
- [x] `--scope` on non-scope target (e.g., opencode) → error
|
||||
- [x] `--also windsurf` uses windsurf's default scope ("global")
|
||||
- [x] `isTargetScope` type guard correctly identifies valid/invalid values
|
||||
|
||||
### Phase 6: Documentation
|
||||
|
||||
**Files:** `README.md`, `CHANGELOG.md`
|
||||
|
||||
- [x] Update README.md Windsurf section to mention `--scope` flag and global default
|
||||
- [x] Add CHANGELOG entry for v0.11.0 with breaking changes documented
|
||||
- [x] Document migration path: `--scope workspace` for old behavior
|
||||
- [x] Note that Windsurf `mcp_config.json` is global-only (workspace MCP config may not be discovered)
|
||||
|
||||
## Acceptance Criteria
|
||||
|
||||
- [x] `install compound-engineering --to windsurf` writes to `~/.codeium/windsurf/` by default
|
||||
- [x] `install compound-engineering --to windsurf --scope workspace` writes to `cwd/.windsurf/`
|
||||
- [x] `--output /custom/path` overrides scope for both commands
|
||||
- [x] `--scope` on non-supporting target produces clear error
|
||||
- [x] `mcp_config.json` merges with existing file (backup created, user entries preserved)
|
||||
- [x] `mcp_config.json` written with `0o600` permissions (not world-readable)
|
||||
- [x] No AGENTS.md generated for either scope
|
||||
- [x] Env var secrets included in `mcp_config.json` with `console.warn` listing affected servers
|
||||
- [x] Both stdio and HTTP/SSE MCP servers included in `mcp_config.json`
|
||||
- [x] All existing tests updated, all new tests pass
|
||||
- [x] No regressions in other targets
|
||||
- [x] `resolveTargetOutputRoot` extracted to shared utility (no duplication)
|
||||
|
||||
## Dependencies & Risks
|
||||
|
||||
**Risk: Global workflow path is undocumented.** Windsurf may not discover workflows from `~/.codeium/windsurf/workflows/`. Mitigation: documented as a known assumption in the brainstorm. Users can `--scope workspace` if global workflows aren't discovered.
|
||||
|
||||
**Risk: Breaking changes for existing v0.10.0 users.** Mitigation: document migration path clearly. `--scope workspace` restores previous behavior. Target is experimental with a small user base.
|
||||
|
||||
**Risk: Workspace `mcp_config.json` not read by Windsurf.** Per Windsurf docs, `mcp_config.json` is global-only configuration. Workspace scope writes the file for forward-compatibility but emits a warning. The primary use case is global scope anyway.
|
||||
|
||||
**Risk: Secrets in `mcp_config.json` committed to git.** Mitigation: `0o600` file permissions, console.warn about sensitive env vars, warning about `.gitignore` for workspace scope.
|
||||
|
||||
## References & Research
|
||||
|
||||
- Spec: `docs/specs/windsurf.md` (authoritative reference for component mapping)
|
||||
- Kiro MCP merge pattern: [src/targets/kiro.ts:68-92](../../src/targets/kiro.ts)
|
||||
- Sync secrets warning: [src/commands/sync.ts:20-28](../../src/commands/sync.ts)
|
||||
- Windsurf MCP docs: https://docs.windsurf.com/windsurf/cascade/mcp
|
||||
- Windsurf Skills global path: https://docs.windsurf.com/windsurf/cascade/skills
|
||||
- Windsurf MCP tutorial: https://windsurf.com/university/tutorials/configuring-first-mcp-server
|
||||
- Adding converter targets (learning): [docs/solutions/adding-converter-target-providers.md](../solutions/adding-converter-target-providers.md)
|
||||
- Plugin versioning (learning): [docs/solutions/plugin-versioning-requirements.md](../solutions/plugin-versioning-requirements.md)
|
||||
@@ -0,0 +1,261 @@
|
||||
---
|
||||
title: "feat: Add ce:* command aliases with backwards-compatible deprecation of workflows:*"
|
||||
type: feat
|
||||
status: active
|
||||
date: 2026-03-01
|
||||
---
|
||||
|
||||
# feat: Add `ce:*` Command Aliases with Backwards-Compatible Deprecation of `workflows:*`
|
||||
|
||||
## Overview
|
||||
|
||||
Rename the five `workflows:*` commands to `ce:*` to make it clearer they belong to compound-engineering. Keep `workflows:*` working as thin deprecation wrappers that warn users and forward to the new commands.
|
||||
|
||||
## Problem Statement / Motivation
|
||||
|
||||
The current `workflows:plan`, `workflows:work`, `workflows:review`, `workflows:brainstorm`, and `workflows:compound` commands are prefixed with `workflows:` — a generic namespace that doesn't signal their origin. Users don't immediately associate them with the compound-engineering plugin.
|
||||
|
||||
The `ce:` prefix is shorter, more memorable, and unambiguously identifies these as compound-engineering commands — consistent with how other plugin commands already use `compound-engineering:` as a namespace.
|
||||
|
||||
## Proposed Solution
|
||||
|
||||
### 1. Create New `ce:*` Commands (Primary)
|
||||
|
||||
Create a `commands/ce/` directory with five new command files. Each file gets the full implementation content from the current `workflows:*` counterpart, with the `name:` frontmatter updated to the new name.
|
||||
|
||||
| New Command | Source Content |
|
||||
|-------------|---------------|
|
||||
| `ce:plan` | `commands/workflows/plan.md` |
|
||||
| `ce:work` | `commands/workflows/work.md` |
|
||||
| `ce:review` | `commands/workflows/review.md` |
|
||||
| `ce:brainstorm` | `commands/workflows/brainstorm.md` |
|
||||
| `ce:compound` | `commands/workflows/compound.md` |
|
||||
|
||||
### 2. Convert `workflows:*` to Deprecation Wrappers (Backwards Compatibility)
|
||||
|
||||
Replace the full content of each `workflows:*` command with a thin wrapper that:
|
||||
1. Displays a visible deprecation warning to the user
|
||||
2. Invokes the new `ce:*` command with the same `$ARGUMENTS`
|
||||
|
||||
Example wrapper body:
|
||||
|
||||
```markdown
|
||||
---
|
||||
name: workflows:plan
|
||||
description: "[DEPRECATED] Use /ce:plan instead. Renamed for clarity."
|
||||
argument-hint: "[feature description]"
|
||||
---
|
||||
|
||||
> ⚠️ **Deprecated:** `/workflows:plan` has been renamed to `/ce:plan`.
|
||||
> Please update your workflow to use `/ce:plan` instead.
|
||||
> This alias will be removed in a future version.
|
||||
|
||||
/ce:plan $ARGUMENTS
|
||||
```
|
||||
|
||||
### 3. Update All Internal References
|
||||
|
||||
The grep reveals `workflows:*` is referenced in **many more places** than just `lfg`/`slfg`. All of these must be updated to point to the new `ce:*` names:
|
||||
|
||||
**Orchestration commands (update to new names):**
|
||||
- `commands/lfg.md` — `/workflows:plan`, `/workflows:work`, `/workflows:review`
|
||||
- `commands/slfg.md` — `/workflows:plan`, `/workflows:work`, `/workflows:review`
|
||||
|
||||
**Command bodies that cross-reference (update to new names):**
|
||||
- `commands/workflows/brainstorm.md` — references `/workflows:plan` multiple times (will be in the deprecated wrapper, so should forward to `/ce:plan`)
|
||||
- `commands/workflows/compound.md` — self-references and references `/workflows:plan`
|
||||
- `commands/workflows/plan.md` — references `/workflows:work` multiple times
|
||||
- `commands/deepen-plan.md` — references `/workflows:work`, `/workflows:compound`
|
||||
|
||||
**Agents (update to new names):**
|
||||
- `agents/review/code-simplicity-reviewer.md` — references `/workflows:plan` and `/workflows:work`
|
||||
- `agents/research/git-history-analyzer.md` — references `/workflows:plan`
|
||||
- `agents/research/learnings-researcher.md` — references `/workflows:plan`
|
||||
|
||||
**Skills (update to new names):**
|
||||
- `skills/document-review/SKILL.md` — references `/workflows:brainstorm`, `/workflows:plan`
|
||||
- `skills/git-worktree/SKILL.md` — references `/workflows:review`, `/workflows:work` extensively
|
||||
- `skills/setup/SKILL.md` — references `/workflows:review`, `/workflows:work`
|
||||
- `skills/brainstorming/SKILL.md` — references `/workflows:plan` multiple times
|
||||
- `skills/file-todos/SKILL.md` — references `/workflows:review`
|
||||
|
||||
**Other commands (update to new names):**
|
||||
- `commands/test-xcode.md` — references `/workflows:review`
|
||||
|
||||
**Historical docs (leave as-is — they document the old names intentionally):**
|
||||
- `docs/plans/*.md` — old plan files, historical record
|
||||
- `docs/brainstorms/*.md` — historical
|
||||
- `docs/solutions/*.md` — historical
|
||||
- `tests/fixtures/` — test fixtures for the converter (intentionally use `workflows:*` to test namespace handling)
|
||||
- `CHANGELOG.md` historical entries — don't rewrite history
|
||||
|
||||
### 4. Update Documentation
|
||||
|
||||
- `CHANGELOG.md` — add new entry documenting the rename and deprecation
|
||||
- `plugins/compound-engineering/README.md` — update command table to list `ce:*` as primary, note `workflows:*` as deprecated aliases
|
||||
- `plugins/compound-engineering/CLAUDE.md` — update command listing and the "Why `workflows:`?" section
|
||||
- Root `README.md` — update the command table (lines 133–136)
|
||||
|
||||
### 5. Converter / bunx Install Script Considerations
|
||||
|
||||
The `bunx` install script (`src/commands/install.ts`) **only writes files, never deletes them**. This has two implications:
|
||||
|
||||
**Now (while deprecated wrappers exist):** No stale file problem. Running `bunx install compound-engineering --to gemini` after this change will:
|
||||
- Write `commands/ce/plan.toml` (new primary)
|
||||
- Write `commands/workflows/plan.toml` (deprecated wrapper, with deprecation content)
|
||||
|
||||
Both coexist correctly. Users who re-run install get both.
|
||||
|
||||
**Future (when deprecated wrappers are eventually removed):** The old `commands/workflows/` files will remain stale in users' converted targets. At that point, a cleanup step will be needed — either:
|
||||
- Manual instructions: "Delete `.gemini/commands/workflows/` after upgrading"
|
||||
- OR add a cleanup pass to the install script that removes known-renamed command directories
|
||||
|
||||
For now, document in the plan that stale cleanup is a known future concern when `workflows:*` wrappers are eventually dropped.
|
||||
|
||||
## Technical Considerations
|
||||
|
||||
### Command Naming
|
||||
|
||||
The `ce:` prefix maps to a `commands/ce/` directory. This follows the existing convention where `workflows:plan` maps to `commands/workflows/plan.md`.
|
||||
|
||||
### Deprecation Warning Display
|
||||
|
||||
Since commands are executed by Claude, the deprecation message in the wrapper body will be displayed to the user as Claude's response before the new command runs. The `>` blockquote markdown renders as a styled callout.
|
||||
|
||||
The deprecated wrappers should **not** use `disable-model-invocation: true` — Claude needs to process the body to display the warning and invoke the new command.
|
||||
|
||||
### Deprecation Wrapper Mechanism
|
||||
|
||||
The deprecated wrappers **must** use `disable-model-invocation: true`. This is the same mechanism `lfg.md` uses — the CLI runtime parses the body and executes slash command invocations directly. Without it, Claude reads the body as text and cannot actually invoke `/ce:plan`.
|
||||
|
||||
The deprecation notice in the wrapper body becomes a printed note (same as `lfg` step descriptions), not a styled Claude response. That's acceptable — it still communicates the message.
|
||||
|
||||
### Context Token Budget
|
||||
|
||||
The 5 new `ce:*` commands add descriptions to the context budget. Keep descriptions short (under 120 chars). The 5 deprecated `workflows:*` wrappers have minimal descriptions (tagged as deprecated) to minimize budget impact.
|
||||
|
||||
### Count Impact
|
||||
|
||||
Command count remains 22 (5 new `ce:*` + 5 updated `workflows:*` wrappers = net zero change). No version bump required for counts.
|
||||
|
||||
## Acceptance Criteria
|
||||
|
||||
- [ ] `commands/ce/` directory created with 5 new command files
|
||||
- [ ] Each `ce:*` command has the full implementation from its `workflows:*` counterpart
|
||||
- [ ] Each `ce:*` command frontmatter `name:` field set to `ce:plan`, `ce:work`, etc.
|
||||
- [ ] Each `workflows:*` command replaced with a thin deprecation wrapper
|
||||
- [ ] Deprecation wrapper shows a clear ⚠️ warning with the new command name
|
||||
- [ ] Deprecation wrapper invokes the new `ce:*` command with `$ARGUMENTS`
|
||||
- [ ] `lfg.md` updated to use `ce:plan`, `ce:work`, `ce:review`
|
||||
- [ ] `slfg.md` updated to use `ce:plan`, `ce:work`, `ce:review`
|
||||
- [ ] All agent `.md` files updated (code-simplicity-reviewer, git-history-analyzer, learnings-researcher)
|
||||
- [ ] All skill `SKILL.md` files updated (document-review, git-worktree, setup, brainstorming, file-todos)
|
||||
- [ ] `commands/deepen-plan.md` and `commands/test-xcode.md` updated
|
||||
- [ ] `CHANGELOG.md` updated with deprecation notice
|
||||
- [ ] `plugins/compound-engineering/README.md` command table updated
|
||||
- [ ] `plugins/compound-engineering/CLAUDE.md` command listing updated
|
||||
- [ ] Root `README.md` command table updated
|
||||
- [ ] Validate: `/ce:plan "test feature"` works end-to-end
|
||||
- [ ] Validate: `/workflows:plan "test feature"` shows deprecation warning and continues
|
||||
- [ ] Re-run `bunx install compound-engineering --to [target]` and confirm both `ce/` and `workflows/` output dirs are written correctly
|
||||
|
||||
## Implementation Steps
|
||||
|
||||
### Step 1: Create `commands/ce/` directory with 5 new files
|
||||
|
||||
For each command, copy the source file and update only the `name:` frontmatter field:
|
||||
|
||||
- `commands/ce/plan.md` — copy `commands/workflows/plan.md`, set `name: ce:plan`
|
||||
- `commands/ce/work.md` — copy `commands/workflows/work.md`, set `name: ce:work`
|
||||
- `commands/ce/review.md` — copy `commands/workflows/review.md`, set `name: ce:review`
|
||||
- `commands/ce/brainstorm.md` — copy `commands/workflows/brainstorm.md`, set `name: ce:brainstorm`
|
||||
- `commands/ce/compound.md` — copy `commands/workflows/compound.md`, set `name: ce:compound`
|
||||
|
||||
### Step 2: Replace `commands/workflows/*.md` with deprecation wrappers
|
||||
|
||||
Use `disable-model-invocation: true` so the CLI runtime directly invokes `/ce:<command>`. The deprecation note is printed as a step description.
|
||||
|
||||
Template for each wrapper:
|
||||
|
||||
```markdown
|
||||
---
|
||||
name: workflows:<command>
|
||||
description: "[DEPRECATED] Use /ce:<command> instead — renamed for clarity."
|
||||
argument-hint: "[...]"
|
||||
disable-model-invocation: true
|
||||
---
|
||||
|
||||
NOTE: /workflows:<command> is deprecated. Please use /ce:<command> instead. This alias will be removed in a future version.
|
||||
|
||||
/ce:<command> $ARGUMENTS
|
||||
```
|
||||
|
||||
### Step 3: Update all internal references
|
||||
|
||||
**Orchestration commands:**
|
||||
- `commands/lfg.md` — replace `/workflows:plan`, `/workflows:work`, `/workflows:review`
|
||||
- `commands/slfg.md` — same
|
||||
|
||||
**Command bodies:**
|
||||
- `commands/deepen-plan.md` — replace `/workflows:work`, `/workflows:compound`
|
||||
- `commands/test-xcode.md` — replace `/workflows:review`
|
||||
- The deprecated `workflows/brainstorm.md`, `workflows/compound.md`, `workflows/plan.md` wrappers — references in their body text pointing to other `workflows:*` commands should also be updated to `ce:*` (since users reading them should see the new names)
|
||||
|
||||
**Agents:**
|
||||
- `agents/review/code-simplicity-reviewer.md`
|
||||
- `agents/research/git-history-analyzer.md`
|
||||
- `agents/research/learnings-researcher.md`
|
||||
|
||||
**Skills:**
|
||||
- `skills/document-review/SKILL.md`
|
||||
- `skills/git-worktree/SKILL.md`
|
||||
- `skills/setup/SKILL.md`
|
||||
- `skills/brainstorming/SKILL.md`
|
||||
- `skills/file-todos/SKILL.md`
|
||||
|
||||
### Step 4: Update documentation
|
||||
|
||||
**`plugins/compound-engineering/CHANGELOG.md`** — Add under new version section:
|
||||
```
|
||||
### Changed
|
||||
- `workflows:plan`, `workflows:work`, `workflows:review`, `workflows:brainstorm`, `workflows:compound` renamed to `ce:plan`, `ce:work`, `ce:review`, `ce:brainstorm`, `ce:compound` for clarity
|
||||
|
||||
### Deprecated
|
||||
- `workflows:*` commands — use `ce:*` equivalents instead. Aliases remain functional and will be removed in a future version.
|
||||
```
|
||||
|
||||
**`plugins/compound-engineering/README.md`** — Update the commands table to list `ce:*` as primary, show `workflows:*` as deprecated aliases.
|
||||
|
||||
**`plugins/compound-engineering/CLAUDE.md`** — Update command listing and the "Why `workflows:`?" section to reflect new `ce:` namespace.
|
||||
|
||||
**Root `README.md`** — Update the commands table (lines 133–136).
|
||||
|
||||
### Step 5: Verify converter output
|
||||
|
||||
After updating, re-run the bunx install script to confirm both targets are written:
|
||||
|
||||
```bash
|
||||
bunx @every-env/compound-plugin install compound-engineering --to gemini --output /tmp/test-output
|
||||
ls /tmp/test-output/.gemini/commands/
|
||||
# Should show both: ce/ and workflows/
|
||||
```
|
||||
|
||||
The `workflows/` output will contain the deprecation wrapper content. The `ce/` output will have the full implementation.
|
||||
|
||||
**Future cleanup note:** When `workflows:*` wrappers are eventually removed, users must manually delete the stale `workflows/` directories from their converted targets (`.gemini/commands/workflows/`, `.codex/commands/workflows/`, etc.). Consider adding a migration note to the CHANGELOG at that time.
|
||||
|
||||
### Step 6: Run `/release-docs` to update the docs site
|
||||
|
||||
## Dependencies & Risks
|
||||
|
||||
- **Risk:** Users with saved references to `workflows:*` commands in their CLAUDE.md files or scripts. **Mitigation:** The deprecation wrappers remain functional indefinitely.
|
||||
- **Risk:** Context token budget slightly increases (5 new command descriptions). **Mitigation:** Keep all descriptions short. Deprecated wrappers get minimal descriptions.
|
||||
- **Risk:** `lfg`/`slfg` orchestration breaks if update is partial. **Mitigation:** Update both in the same commit.
|
||||
|
||||
## Sources & References
|
||||
|
||||
- Existing commands: `plugins/compound-engineering/commands/workflows/*.md`
|
||||
- Orchestration commands: `plugins/compound-engineering/commands/lfg.md`, `plugins/compound-engineering/commands/slfg.md`
|
||||
- Plugin metadata: `plugins/compound-engineering/.claude-plugin/plugin.json`
|
||||
- Changelog: `plugins/compound-engineering/CHANGELOG.md`
|
||||
- README: `plugins/compound-engineering/README.md`
|
||||
@@ -0,0 +1,140 @@
|
||||
---
|
||||
title: "fix: Setup skill fails silently on non-Claude LLMs due to AskUserQuestion dependency"
|
||||
type: fix
|
||||
status: active
|
||||
date: 2026-03-01
|
||||
---
|
||||
|
||||
## Enhancement Summary
|
||||
|
||||
**Deepened on:** 2026-03-01
|
||||
**Research agents used:** best-practices-researcher, architecture-strategist, code-simplicity-reviewer, scope-explorer
|
||||
|
||||
### Key Improvements
|
||||
1. Simplified preamble from 16 lines to 4 lines — drop platform name list and example blockquote (YAGNI)
|
||||
2. Expanded scope: `create-new-skill.md` also has `AskUserQuestion` and needs the same fix
|
||||
3. Clarified that `codex-agents.ts` change helps command/agent contexts only — does NOT reach skill execution (skills aren't converter-transformed)
|
||||
4. Added CLAUDE.md skill compliance policy as a third deliverable to prevent recurrence
|
||||
5. Separated two distinct failure modes: tool-not-found error vs silent auto-configuration
|
||||
|
||||
### New Considerations Discovered
|
||||
- Only Pi converter transforms `AskUserQuestion` (incompletely); all others pass skill content through verbatim — the codex-agents.ts fix is independent of skill execution
|
||||
- `add-workflow.md` and `audit-skill.md` already explicitly prohibit `AskUserQuestion` — this undocumented policy should be formalized
|
||||
- Prose fallback is probabilistic (LLM compliance); converter-level transformation is the correct long-term architectural fix
|
||||
- The brainstorming skill avoids `AskUserQuestion` entirely and works cross-platform — that's the gold standard pattern
|
||||
|
||||
---
|
||||
|
||||
# fix: Setup Skill Cross-Platform Fallback for AskUserQuestion
|
||||
|
||||
## Overview
|
||||
|
||||
The `setup` skill uses `AskUserQuestion` at 5 decision points. On non-Claude platforms (Codex, Gemini, OpenCode, Copilot, Kiro, etc.), this tool doesn't exist — the LLM reads the skill body but cannot call the tool, causing silent failure or unconsented auto-configuration. Fix by adding a minimal fallback instruction to the skill body, applying the same to `create-new-skill.md`, and adding a policy to the CLAUDE.md skill checklist to prevent recurrence.
|
||||
|
||||
## Problem Statement
|
||||
|
||||
**Two distinct failure modes:**
|
||||
|
||||
1. **Tool-not-found error** — LLM tries to call `AskUserQuestion` as a function; platform returns an error. Setup halts.
|
||||
2. **Silent skip** — LLM reads `AskUserQuestion` as prose, ignores the decision gate, auto-configures. User never consulted. This is worse — produces a `compound-engineering.local.md` the user never approved.
|
||||
|
||||
`plugins/compound-engineering/skills/setup/SKILL.md` has 5 `AskUserQuestion` blocks:
|
||||
|
||||
| Line | Decision Point |
|
||||
|------|----------------|
|
||||
| 13 | Check existing config: Reconfigure / View / Cancel |
|
||||
| 44 | Stack detection: Auto-configure / Customize |
|
||||
| 67 | Stack override (multi-option) |
|
||||
| 85 | Focus areas (multiSelect) |
|
||||
| 104 | Review depth: Thorough / Fast / Comprehensive |
|
||||
|
||||
`plugins/compound-engineering/skills/create-agent-skills/workflows/create-new-skill.md` lines 22 and 45 also use `AskUserQuestion`.
|
||||
|
||||
Only the Pi converter transforms the reference (incompletely). All other converters (Codex, Gemini, Copilot, Kiro, Droid, Windsurf) pass skill content through verbatim — **skills are not converter-transformed**.
|
||||
|
||||
## Proposed Solution
|
||||
|
||||
Three deliverables, each addressing a different layer:
|
||||
|
||||
### 1. Add 4-line "Interaction Method" preamble to `setup/SKILL.md`
|
||||
|
||||
Immediately after the `# Compound Engineering Setup` heading, insert:
|
||||
|
||||
```markdown
|
||||
## Interaction Method
|
||||
|
||||
If `AskUserQuestion` is available, use it for all prompts below.
|
||||
|
||||
If not, present each question as a numbered list and wait for a reply before proceeding to the next step. For multiSelect questions, accept comma-separated numbers (e.g. `1, 3`). Never skip or auto-configure.
|
||||
```
|
||||
|
||||
**Why 4 lines, not 16:** LLMs know what a numbered list is — no example blockquote needed. The branching condition is tool availability, not platform identity — no platform name list needed (YAGNI: new platforms will be added and lists go stale). State the "never skip" rule once here; don't repeat it in `codex-agents.ts`.
|
||||
|
||||
**Why this works:** The skill body IS read by the LLM on all platforms when `/setup` is invoked. The agent follows prose instructions regardless of tool availability. This is the same pattern `brainstorming/SKILL.md` uses — it avoids `AskUserQuestion` entirely and uses inline numbered lists — the gold standard cross-platform approach.
|
||||
|
||||
### 2. Apply the same preamble to `create-new-skill.md`
|
||||
|
||||
`plugins/compound-engineering/skills/create-agent-skills/workflows/create-new-skill.md` uses `AskUserQuestion` at lines 22 and 45. Apply an identical preamble at the top of that file.
|
||||
|
||||
### 3. Strengthen `codex-agents.ts` AskUserQuestion mapping
|
||||
|
||||
This change does NOT fix skill execution (skills bypass the converter pipeline). It improves the AGENTS.md guidance for Codex command/agent contexts.
|
||||
|
||||
Replace (`src/utils/codex-agents.ts` line 21):
|
||||
```
|
||||
- AskUserQuestion/Question: ask the user in chat
|
||||
```
|
||||
|
||||
With:
|
||||
```
|
||||
- AskUserQuestion/Question: present choices as a numbered list in chat and wait for a reply number. For multi-select (multiSelect: true), accept comma-separated numbers. Never skip or auto-configure — always wait for the user's response before proceeding.
|
||||
```
|
||||
|
||||
### 4. Add lint rule to CLAUDE.md skill compliance checklist
|
||||
|
||||
Add to the "Skill Compliance Checklist" in `plugins/compound-engineering/CLAUDE.md`:
|
||||
|
||||
```
|
||||
### AskUserQuestion Usage
|
||||
|
||||
- [ ] If the skill uses `AskUserQuestion`, it must include an "Interaction Method" preamble explaining the numbered-list fallback for non-Claude environments
|
||||
- [ ] Prefer avoiding `AskUserQuestion` entirely (see brainstorming/SKILL.md pattern) for skills intended to run cross-platform
|
||||
```
|
||||
|
||||
## Technical Considerations
|
||||
|
||||
- `setup/SKILL.md` has `disable-model-invocation: true` — this controls session-startup context loading only, not skill-body execution at invocation time
|
||||
- The prose fallback is probabilistic (LLM compliance), not a build-time guarantee. The correct long-term architectural fix is converter-level transformation of skill content (a `transformSkillContent()` pass in each converter), but that is out of scope here
|
||||
- Commands with `AskUserQuestion` (`ce/brainstorm.md`, `ce/plan.md`, `test-browser.md`, etc.) have the same gap but are out of scope — explicitly noted as a future task
|
||||
|
||||
## Acceptance Criteria
|
||||
|
||||
- [ ] `setup/SKILL.md` has a 4-line "Interaction Method" preamble after the opening heading
|
||||
- [ ] `create-new-skill.md` has the same preamble
|
||||
- [ ] The skills still use `AskUserQuestion` as primary — no change to Claude Code behavior
|
||||
- [ ] `codex-agents.ts` AskUserQuestion line updated with structured guidance
|
||||
- [ ] `plugins/compound-engineering/CLAUDE.md` skill checklist includes AskUserQuestion policy
|
||||
- [ ] No regression: on Claude Code, setup works exactly as before
|
||||
|
||||
## Files
|
||||
|
||||
- `plugins/compound-engineering/skills/setup/SKILL.md` — Add 4-line preamble after line 8
|
||||
- `plugins/compound-engineering/skills/create-agent-skills/workflows/create-new-skill.md` — Add same preamble at top
|
||||
- `src/utils/codex-agents.ts` — Strengthen AskUserQuestion mapping (line 21)
|
||||
- `plugins/compound-engineering/CLAUDE.md` — Add AskUserQuestion policy to skill compliance checklist
|
||||
|
||||
## Future Work (Out of Scope)
|
||||
|
||||
- Converter-level `transformSkillContent()` for all targets — build-time guarantee instead of prose fallback
|
||||
- Commands with `AskUserQuestion` (`ce/brainstorm.md`, `ce/plan.md`, `test-browser.md`) — same failure mode, separate fix
|
||||
|
||||
## Sources & References
|
||||
|
||||
- Issue: [#204](https://github.com/EveryInc/compound-engineering-plugin/issues/204)
|
||||
- `plugins/compound-engineering/skills/setup/SKILL.md:13,44,67,85,104`
|
||||
- `plugins/compound-engineering/skills/create-agent-skills/workflows/create-new-skill.md:22,45`
|
||||
- `src/utils/codex-agents.ts:21`
|
||||
- `src/converters/claude-to-pi.ts:106` — Pi converter (reference pattern)
|
||||
- `plugins/compound-engineering/skills/brainstorming/SKILL.md` — gold standard cross-platform skill (no AskUserQuestion)
|
||||
- `plugins/compound-engineering/skills/create-agent-skills/workflows/add-workflow.md:12,37` — existing "DO NOT use AskUserQuestion" policy
|
||||
- `docs/solutions/adding-converter-target-providers.md`
|
||||
@@ -0,0 +1,639 @@
|
||||
---
|
||||
title: "feat: Sync Claude MCP servers to all supported providers"
|
||||
type: feat
|
||||
date: 2026-03-03
|
||||
status: completed
|
||||
deepened: 2026-03-03
|
||||
---
|
||||
|
||||
# feat: Sync Claude MCP servers to all supported providers
|
||||
|
||||
## Overview
|
||||
|
||||
Expand the `sync` command so a user's local Claude Code MCP configuration can be propagated to every provider this CLI can reasonably support, instead of only the current partial set.
|
||||
|
||||
Today, `sync` already symlinks Claude skills and syncs MCP servers for a subset of targets. The gap is that install/convert support has grown much faster than sync support, so the product promise in `README.md` has drifted away from what `src/commands/sync.ts` can actually do.
|
||||
|
||||
This feature should close that parity gap without changing the core sync contract:
|
||||
|
||||
- Claude remains the source of truth for personal skills and MCP servers.
|
||||
- Skills stay symlinked, not copied.
|
||||
- Existing user config in the destination tool is preserved where possible.
|
||||
- Target-specific MCP formats stay target-specific.
|
||||
|
||||
## Problem Statement
|
||||
|
||||
The current implementation has three concrete problems:
|
||||
|
||||
1. `sync` only knows about `opencode`, `codex`, `pi`, `droid`, `copilot`, and `gemini`, while install/convert now supports `kiro`, `windsurf`, `openclaw`, and `qwen` too.
|
||||
2. `sync --target all` relies on stale detection metadata that still includes `cursor`, but misses newer supported tools.
|
||||
3. Existing MCP sync support is incomplete even for some already-supported targets:
|
||||
- `codex` only emits stdio servers and silently drops remote MCP servers.
|
||||
- `droid` is still skills-only even though Factory now documents `mcp.json`.
|
||||
|
||||
User impact:
|
||||
|
||||
- A user can install the plugin to more providers than they can sync their personal Claude setup to.
|
||||
- `sync --target all` does not mean "all supported tools" anymore.
|
||||
- Users with remote MCP servers in Claude get partial results depending on target.
|
||||
|
||||
## Research Summary
|
||||
|
||||
### No Relevant Brainstorm
|
||||
|
||||
I checked recent brainstorms in `docs/brainstorms/` and found no relevant document for this feature within the last 14 days.
|
||||
|
||||
### Internal Findings
|
||||
|
||||
- `src/commands/sync.ts:15-125` hardcodes the sync target list, output roots, and per-target dispatch. It omits `windsurf`, `kiro`, `openclaw`, and `qwen`.
|
||||
- `src/utils/detect-tools.ts:15-22` still detects `cursor`, but not `windsurf`, `kiro`, `openclaw`, or `qwen`.
|
||||
- `src/parsers/claude-home.ts:11-19` already gives sync exactly the right inputs: personal skills plus `settings.json` `mcpServers`.
|
||||
- `src/sync/codex.ts:25-91` only serializes stdio MCP servers, even though Codex supports remote MCP config.
|
||||
- `src/sync/droid.ts:6-21` symlinks skills but ignores MCP entirely.
|
||||
- Target writers already encode several missing MCP formats and merge behaviors:
|
||||
- `src/targets/windsurf.ts:65-92`
|
||||
- `src/targets/kiro.ts:68-91`
|
||||
- `src/targets/openclaw.ts:34-42`
|
||||
- `src/targets/qwen.ts:9-15`
|
||||
- `README.md:89-123` promises "Sync Personal Config" but only documents the old subset of targets.
|
||||
|
||||
### Institutional Learnings
|
||||
|
||||
`docs/solutions/adding-converter-target-providers.md:20-32` and `docs/solutions/adding-converter-target-providers.md:208-214` reinforce the right pattern for this feature:
|
||||
|
||||
- keep target mappings explicit,
|
||||
- treat MCP conversion as target-specific,
|
||||
- warn on unsupported features instead of forcing fake parity,
|
||||
- and add tests for each mapping.
|
||||
|
||||
Note: `docs/solutions/patterns/critical-patterns.md` does not exist in this repository, so there was no critical-patterns file to apply.
|
||||
|
||||
### External Findings
|
||||
|
||||
Official docs confirm that the missing targets are not all equivalent, so this cannot be solved with a generic JSON pass-through.
|
||||
|
||||
| Target | Official MCP / skills location | Key notes |
|
||||
| --- | --- | --- |
|
||||
| Factory Droid | `~/.factory/mcp.json`, `.factory/mcp.json`, `~/.factory/skills/` | Supports `stdio` and `http`; user config overrides project config. |
|
||||
| Windsurf | `~/.codeium/windsurf/mcp_config.json`, `~/.codeium/windsurf/skills/` | Supports `stdio`, Streamable HTTP, and SSE; remote config uses `serverUrl` or `url`. |
|
||||
| Kiro | `~/.kiro/settings/mcp.json`, `.kiro/settings/mcp.json`, `~/.kiro/skills/` | Supports user and workspace config; remote MCP support was added after this repo's local Kiro spec was written. |
|
||||
| Qwen Code | `~/.qwen/settings.json`, `.qwen/settings.json`, `~/.qwen/skills/`, `.qwen/skills/` | Supports `stdio`, `http`, and `sse`; official docs say prefer `http`, with `sse` treated as legacy/deprecated. |
|
||||
| OpenClaw | `~/.openclaw/skills`, `<workspace>/skills`, `~/.openclaw/openclaw.json` | Skills are well-documented; a generic MCP server config surface is not clearly documented in official docs, so MCP sync needs validation before implementation is promised. |
|
||||
|
||||
Additional important findings:
|
||||
|
||||
- Kiro's current official behavior supersedes the local repo spec that says "workspace only" and "stdio only".
|
||||
- Qwen's current docs explicitly distinguish `httpUrl` from legacy SSE `url`; blindly copying Claude's `url` is too lossy.
|
||||
- Factory and Windsurf both support remote MCP, so `droid` should no longer be treated as skills-only.
|
||||
|
||||
## Proposed Solution
|
||||
|
||||
### Product Decision
|
||||
|
||||
Treat this as **sync parity for MCP-capable providers**, not as a one-off patch.
|
||||
|
||||
That means this feature should:
|
||||
|
||||
- add missing sync targets where the provider has a documented skills/MCP surface,
|
||||
- upgrade partial implementations where existing sync support drops valid Claude MCP data,
|
||||
- and replace stale detection metadata so `sync --target all` is truthful again.
|
||||
|
||||
### Scope
|
||||
|
||||
#### In Scope
|
||||
|
||||
- Add MCP sync coverage for:
|
||||
- `droid`
|
||||
- `windsurf`
|
||||
- `kiro`
|
||||
- `qwen`
|
||||
- Expand `codex` sync to support remote MCP servers.
|
||||
- Add provider detection for newly supported sync targets.
|
||||
- Keep skills syncing for all synced targets.
|
||||
- Update CLI help text, README sync docs, and tests.
|
||||
|
||||
#### Conditional / Validation Gate
|
||||
|
||||
- `openclaw` skills sync is straightforward and should be included if the target is added to `sync`.
|
||||
- `openclaw` MCP sync should only be implemented if its config surface is validated against current upstream docs or current upstream source. If that validation fails, the feature should explicitly skip OpenClaw MCP sync with a warning rather than inventing a format.
|
||||
|
||||
#### Out of Scope
|
||||
|
||||
- Standardizing all existing sync targets onto user-level paths only.
|
||||
- Reworking install/convert output roots.
|
||||
- Hook sync.
|
||||
- A full rewrite of target writers.
|
||||
|
||||
### Design Decisions
|
||||
|
||||
#### 0. Keep existing sync roots stable unless this feature is explicitly adding a new target
|
||||
|
||||
Do not use this feature to migrate existing `copilot` and `gemini` sync behavior.
|
||||
|
||||
Backward-compatibility rule:
|
||||
|
||||
- existing targets keep their current sync roots unless a correctness bug forces a change,
|
||||
- newly added sync targets use the provider's documented personal/global config surface,
|
||||
- and any future root migration belongs in a separate plan.
|
||||
|
||||
Planned sync roots after this feature:
|
||||
|
||||
| Target | Sync root | Notes |
|
||||
| --- | --- | --- |
|
||||
| `opencode` | `~/.config/opencode` | unchanged |
|
||||
| `codex` | `~/.codex` | unchanged |
|
||||
| `pi` | `~/.pi/agent` | unchanged |
|
||||
| `droid` | `~/.factory` | unchanged root, new MCP file |
|
||||
| `copilot` | `.github` | unchanged for backwards compatibility |
|
||||
| `gemini` | `.gemini` | unchanged for backwards compatibility |
|
||||
| `windsurf` | `~/.codeium/windsurf` | new |
|
||||
| `kiro` | `~/.kiro` | new |
|
||||
| `qwen` | `~/.qwen` | new |
|
||||
| `openclaw` | `~/.openclaw` | new, MCP still validation-gated |
|
||||
|
||||
#### 1. Add a dedicated sync target registry
|
||||
|
||||
Do not keep growing `sync.ts` as a hand-maintained switch statement.
|
||||
|
||||
Create a dedicated sync registry, for example:
|
||||
|
||||
### `src/sync/registry.ts`
|
||||
|
||||
```ts
|
||||
import os from "os"
|
||||
import path from "path"
|
||||
import type { ClaudeHomeConfig } from "../parsers/claude-home"
|
||||
|
||||
export type SyncTargetDefinition = {
|
||||
name: string
|
||||
detectPaths: (home: string, cwd: string) => string[]
|
||||
resolveOutputRoot: (home: string, cwd: string) => string
|
||||
sync: (config: ClaudeHomeConfig, outputRoot: string) => Promise<void>
|
||||
}
|
||||
```
|
||||
|
||||
This registry becomes the single source of truth for:
|
||||
|
||||
- valid `sync` targets,
|
||||
- `sync --target all` detection,
|
||||
- output root resolution,
|
||||
- and dispatch.
|
||||
|
||||
This avoids the current drift between:
|
||||
|
||||
- `src/commands/sync.ts`
|
||||
- `src/utils/detect-tools.ts`
|
||||
- `README.md`
|
||||
|
||||
#### 2. Preserve sync semantics, not writer semantics
|
||||
|
||||
Do not directly reuse install target writers for sync.
|
||||
|
||||
Reason:
|
||||
|
||||
- writers mostly copy skill directories,
|
||||
- sync intentionally symlinks skills,
|
||||
- writers often emit full plugin/install bundles,
|
||||
- sync only needs personal skills plus MCP config.
|
||||
|
||||
However, provider-specific MCP conversion helpers should be extracted or reused where practical so sync and writer logic do not diverge again.
|
||||
|
||||
#### 3. Keep merge behavior additive, with Claude winning on same-name collisions
|
||||
|
||||
For JSON-based targets:
|
||||
|
||||
- preserve unrelated user keys,
|
||||
- preserve unrelated user MCP servers,
|
||||
- but if the same server name exists in Claude and the target config, Claude's value should overwrite that server entry during sync.
|
||||
|
||||
Codex remains the special case:
|
||||
|
||||
- continue using the managed marker block,
|
||||
- remove the previous managed block,
|
||||
- rewrite the managed block from Claude,
|
||||
- leave the rest of `config.toml` untouched.
|
||||
|
||||
#### 4. Secure config writes where secrets may exist
|
||||
|
||||
Any config file that may contain MCP headers or env vars should be written with restrictive permissions where the platform already supports that pattern.
|
||||
|
||||
At minimum:
|
||||
|
||||
- `config.toml`
|
||||
- `mcp.json`
|
||||
- `mcp_config.json`
|
||||
- `settings.json`
|
||||
|
||||
should follow the repo's existing "secure write" conventions where possible.
|
||||
|
||||
#### 5. Do not silently coerce ambiguous remote transports
|
||||
|
||||
Qwen and possibly future targets distinguish Streamable HTTP from legacy SSE.
|
||||
|
||||
Use this mapping rule:
|
||||
|
||||
- if Claude explicitly provides `type: "sse"` or an equivalent known signal, map to the target's SSE field,
|
||||
- otherwise prefer the target's HTTP form for remote URLs,
|
||||
- and log a warning when a target requires more specificity than Claude provides.
|
||||
|
||||
## Provider Mapping Plan
|
||||
|
||||
### Existing Targets to Upgrade
|
||||
|
||||
#### Codex
|
||||
|
||||
Current issue:
|
||||
|
||||
- only stdio servers are synced.
|
||||
|
||||
Implementation:
|
||||
|
||||
- extend `syncToCodex()` so remote MCP servers are serialized into the Codex TOML format, not dropped.
|
||||
- keep the existing marker-based idempotent section handling.
|
||||
|
||||
Notes:
|
||||
|
||||
- This is a correctness fix, not a new target.
|
||||
|
||||
#### Droid / Factory
|
||||
|
||||
Current issue:
|
||||
|
||||
- skills-only sync despite current official MCP support.
|
||||
|
||||
Implementation:
|
||||
|
||||
- add `src/sync/droid.ts` MCP config writing to `~/.factory/mcp.json`.
|
||||
- merge with existing `mcpServers`.
|
||||
- support both `stdio` and `http`.
|
||||
|
||||
### New Sync Targets
|
||||
|
||||
#### Windsurf
|
||||
|
||||
Add `src/sync/windsurf.ts`:
|
||||
|
||||
- symlink Claude skills into `~/.codeium/windsurf/skills/`
|
||||
- merge MCP servers into `~/.codeium/windsurf/mcp_config.json`
|
||||
- support `stdio`, Streamable HTTP, and SSE
|
||||
- prefer `serverUrl` for remote HTTP config
|
||||
- preserve unrelated existing servers
|
||||
- write with secure permissions
|
||||
|
||||
Reference implementation:
|
||||
|
||||
- `src/targets/windsurf.ts:65-92`
|
||||
|
||||
#### Kiro
|
||||
|
||||
Add `src/sync/kiro.ts`:
|
||||
|
||||
- symlink Claude skills into `~/.kiro/skills/`
|
||||
- merge MCP servers into `~/.kiro/settings/mcp.json`
|
||||
- support both local and remote MCP servers
|
||||
- preserve user config already present in `mcp.json`
|
||||
|
||||
Important:
|
||||
|
||||
- This feature must treat the repository's local Kiro spec as stale where it conflicts with official 2025-2026 Kiro docs/blog posts.
|
||||
|
||||
Reference implementation:
|
||||
|
||||
- `src/targets/kiro.ts:68-91`
|
||||
|
||||
#### Qwen
|
||||
|
||||
Add `src/sync/qwen.ts`:
|
||||
|
||||
- symlink Claude skills into `~/.qwen/skills/`
|
||||
- merge MCP servers into `~/.qwen/settings.json`
|
||||
- map stdio directly
|
||||
- map remote URLs to `httpUrl` by default
|
||||
- only emit legacy SSE `url` when Claude transport clearly indicates SSE
|
||||
|
||||
Important:
|
||||
|
||||
- capture the deprecation note in docs/comments: SSE is legacy, so HTTP is the default remote mapping.
|
||||
|
||||
#### OpenClaw
|
||||
|
||||
Add `src/sync/openclaw.ts` only if validated during implementation:
|
||||
|
||||
- symlink skills into `~/.openclaw/skills`
|
||||
- optionally merge MCP config into `~/.openclaw/openclaw.json` if the official/current upstream contract is confirmed
|
||||
|
||||
Fallback behavior if MCP config cannot be validated:
|
||||
|
||||
- sync skills only,
|
||||
- emit a warning that OpenClaw MCP sync is skipped because the official config surface is not documented clearly enough.
|
||||
|
||||
## Implementation Phases
|
||||
|
||||
### Phase 1: Registry and shared helpers
|
||||
|
||||
Files:
|
||||
|
||||
- `src/commands/sync.ts`
|
||||
- `src/utils/detect-tools.ts`
|
||||
- `src/sync/registry.ts` (new)
|
||||
- `src/sync/skills.ts` or `src/utils/symlink.ts` extension
|
||||
- optional `src/sync/mcp-merge.ts`
|
||||
|
||||
Tasks:
|
||||
|
||||
- move sync target metadata into a single registry
|
||||
- make `validTargets` derive from the registry
|
||||
- make `sync --target all` use the registry
|
||||
- update detection to include supported sync targets instead of stale `cursor`
|
||||
- extract a shared helper for validated skill symlinking
|
||||
|
||||
### Phase 2: Upgrade existing partial targets
|
||||
|
||||
Files:
|
||||
|
||||
- `src/sync/codex.ts`
|
||||
- `src/sync/droid.ts`
|
||||
- `tests/sync-droid.test.ts`
|
||||
- new or expanded `tests/sync-codex.test.ts`
|
||||
|
||||
Tasks:
|
||||
|
||||
- add remote MCP support to Codex sync
|
||||
- add MCP config writing to Droid sync
|
||||
- preserve current skill symlink behavior
|
||||
|
||||
### Phase 3: Add missing sync targets
|
||||
|
||||
Files:
|
||||
|
||||
- `src/sync/windsurf.ts`
|
||||
- `src/sync/kiro.ts`
|
||||
- `src/sync/qwen.ts`
|
||||
- optionally `src/sync/openclaw.ts`
|
||||
- `tests/sync-windsurf.test.ts`
|
||||
- `tests/sync-kiro.test.ts`
|
||||
- `tests/sync-qwen.test.ts`
|
||||
- optionally `tests/sync-openclaw.test.ts`
|
||||
|
||||
Tasks:
|
||||
|
||||
- implement skill symlink + MCP merge for each target
|
||||
- align output paths with the target's documented personal config surface
|
||||
- secure writes and corrupted-config fallbacks
|
||||
|
||||
### Phase 4: CLI, docs, and detection parity
|
||||
|
||||
Files:
|
||||
|
||||
- `src/commands/sync.ts`
|
||||
- `src/utils/detect-tools.ts`
|
||||
- `tests/detect-tools.test.ts`
|
||||
- `tests/cli.test.ts`
|
||||
- `README.md`
|
||||
- optionally `docs/specs/kiro.md`
|
||||
|
||||
Tasks:
|
||||
|
||||
- update `sync` help text and summary output
|
||||
- ensure `sync --target all` only reports real sync-capable tools
|
||||
- document newly supported sync targets
|
||||
- fix stale Kiro assumptions if repository docs are updated in the same change
|
||||
|
||||
## SpecFlow Analysis
|
||||
|
||||
### Primary user flows
|
||||
|
||||
#### Flow 1: Explicit sync to one target
|
||||
|
||||
1. User runs `bunx @every-env/compound-plugin sync --target <provider>`
|
||||
2. CLI loads `~/.claude/skills` and `~/.claude/settings.json`
|
||||
3. CLI resolves that provider's sync root
|
||||
4. Skills are symlinked
|
||||
5. MCP config is merged
|
||||
6. CLI prints the destination path and completion summary
|
||||
|
||||
#### Flow 2: Sync to all detected tools
|
||||
|
||||
1. User runs `bunx @every-env/compound-plugin sync`
|
||||
2. CLI detects installed/supported tools
|
||||
3. CLI prints which tools were found and which were skipped
|
||||
4. CLI syncs each detected target in sequence
|
||||
5. CLI prints per-target success lines
|
||||
|
||||
#### Flow 3: Existing config already present
|
||||
|
||||
1. User already has destination config file(s)
|
||||
2. Sync reads and parses the existing file
|
||||
3. Existing unrelated keys are preserved
|
||||
4. Claude MCP entries are merged in
|
||||
5. Corrupt config produces a warning and replacement behavior
|
||||
|
||||
### Edge cases to account for
|
||||
|
||||
- Claude has zero MCP servers: skills still sync, no config file is written.
|
||||
- Claude has remote MCP servers: targets that support remote config receive them; unsupported transports warn, not crash.
|
||||
- Existing target config is invalid JSON/TOML: warn and replace the managed portion.
|
||||
- Skill name contains path traversal characters: skip with warning, same as current behavior.
|
||||
- Real directory already exists where a symlink would go: skip safely, do not delete user data.
|
||||
- `sync --target all` detects a tool with skills support but unclear MCP support: sync only the documented subset and warn explicitly.
|
||||
|
||||
### Critical product decisions already assumed
|
||||
|
||||
- `sync` remains additive and non-destructive.
|
||||
- Sync roots may differ from install roots when the provider has a documented personal config location.
|
||||
- OpenClaw MCP support is validation-gated rather than assumed.
|
||||
|
||||
## Acceptance Criteria
|
||||
|
||||
### Functional Requirements
|
||||
|
||||
- [x] `sync --target` accepts `windsurf`, `kiro`, and `qwen`, in addition to the existing targets.
|
||||
- [x] `sync --target droid` writes MCP servers to Factory's documented `mcp.json` format instead of remaining skills-only.
|
||||
- [x] `sync --target codex` syncs both stdio and remote MCP servers.
|
||||
- [x] `sync --target all` detects only sync-capable supported tools and includes the new targets.
|
||||
- [x] Claude personal skills continue to be symlinked, not copied.
|
||||
- [x] Existing destination config keys unrelated to MCP are preserved during merge.
|
||||
- [x] Existing same-named MCP entries are refreshed from Claude for sync-managed targets.
|
||||
- [x] Unsafe skill names are skipped without deleting user content.
|
||||
- [x] If OpenClaw MCP sync is not validated, the CLI warns and skips MCP sync for OpenClaw instead of writing an invented format.
|
||||
|
||||
### Non-Functional Requirements
|
||||
|
||||
- [x] MCP config files that may contain secrets are written with restrictive permissions where supported.
|
||||
- [x] Corrupt destination config files warn and recover cleanly.
|
||||
- [x] New sync code does not duplicate target detection metadata in multiple places.
|
||||
- [x] Remote transport mapping is explicit and tested, especially for Qwen and Codex.
|
||||
|
||||
### Quality Gates
|
||||
|
||||
- [x] Add target-level sync tests for every new or upgraded provider.
|
||||
- [x] Update `tests/detect-tools.test.ts` for new detection rules and remove stale cursor expectations.
|
||||
- [x] Add or expand CLI coverage for `sync --target all`.
|
||||
- [x] `bun test` passes.
|
||||
|
||||
## Testing Plan
|
||||
|
||||
### Unit / integration tests
|
||||
|
||||
Add or expand:
|
||||
|
||||
- `tests/sync-codex.test.ts`
|
||||
- remote URL server is emitted
|
||||
- existing non-managed TOML content is preserved
|
||||
- `tests/sync-droid.test.ts`
|
||||
- writes `mcp.json`
|
||||
- merges with existing file
|
||||
- `tests/sync-windsurf.test.ts`
|
||||
- writes `mcp_config.json`
|
||||
- merges existing servers
|
||||
- preserves HTTP/SSE fields
|
||||
- `tests/sync-kiro.test.ts`
|
||||
- writes `settings/mcp.json`
|
||||
- supports user-scope root
|
||||
- preserves remote servers
|
||||
- `tests/sync-qwen.test.ts`
|
||||
- writes `settings.json`
|
||||
- maps remote servers to `httpUrl`
|
||||
- emits legacy SSE only when explicitly indicated
|
||||
- `tests/sync-openclaw.test.ts` if implemented
|
||||
- skills path
|
||||
- MCP behavior or explicit skip warning
|
||||
|
||||
### CLI tests
|
||||
|
||||
Expand `tests/cli.test.ts` or add focused sync CLI coverage for:
|
||||
|
||||
- `sync --target windsurf`
|
||||
- `sync --target kiro`
|
||||
- `sync --target qwen`
|
||||
- `sync --target all` with detected new tool homes
|
||||
- `sync --target all` no longer surfacing unsupported `cursor`
|
||||
|
||||
## Risks and Mitigations
|
||||
|
||||
### Risk: local specs are stale relative to current provider docs
|
||||
|
||||
Impact:
|
||||
|
||||
- implementing from local docs alone would produce incorrect paths and transport support.
|
||||
|
||||
Mitigation:
|
||||
|
||||
- treat official 2025-2026 docs/blog posts as source of truth where they supersede local specs
|
||||
- update any obviously stale repo docs touched by this feature
|
||||
|
||||
### Risk: transport ambiguity for remote MCP servers
|
||||
|
||||
Impact:
|
||||
|
||||
- a Claude `url` may map incorrectly for targets that distinguish HTTP vs SSE.
|
||||
|
||||
Mitigation:
|
||||
|
||||
- prefer HTTP where the target recommends it
|
||||
- only emit legacy SSE when Claude transport is explicit
|
||||
- warn when mapping is lossy
|
||||
|
||||
### Risk: OpenClaw MCP surface is not sufficiently documented
|
||||
|
||||
Impact:
|
||||
|
||||
- writing a guessed MCP config could create a broken or misleading feature.
|
||||
|
||||
Mitigation:
|
||||
|
||||
- validation gate during implementation
|
||||
- if validation fails, ship OpenClaw skills sync only and document MCP as a follow-up
|
||||
|
||||
### Risk: `sync --target all` remains easy to drift out of sync again
|
||||
|
||||
Impact:
|
||||
|
||||
- future providers get added to install/convert but missed by sync.
|
||||
|
||||
Mitigation:
|
||||
|
||||
- derive sync valid targets and detection from a shared registry
|
||||
- add tests that assert detection and sync target lists match expected supported names
|
||||
|
||||
## Alternative Approaches Considered
|
||||
|
||||
### 1. Just add more cases to `sync.ts`
|
||||
|
||||
Rejected:
|
||||
|
||||
- this is exactly how the current drift happened.
|
||||
|
||||
### 2. Reuse target writers directly
|
||||
|
||||
Rejected:
|
||||
|
||||
- writers copy directories and emit install bundles;
|
||||
- sync must symlink skills and only manage personal config subsets.
|
||||
|
||||
### 3. Standardize every sync target on user-level output now
|
||||
|
||||
Rejected for this feature:
|
||||
|
||||
- it would change existing `gemini` and `copilot` behavior and broaden scope into a migration project.
|
||||
|
||||
## Documentation Plan
|
||||
|
||||
- Update `README.md` sync section to list all supported sync targets and call out any exceptions.
|
||||
- Update sync examples for `windsurf`, `kiro`, and `qwen`.
|
||||
- If OpenClaw MCP is skipped, document that explicitly.
|
||||
- If repository specs are corrected during implementation, update `docs/specs/kiro.md` to match official current behavior.
|
||||
|
||||
## Success Metrics
|
||||
|
||||
- `sync --target all` covers the same provider surface users reasonably expect from the current CLI, excluding only targets that lack a validated MCP config contract.
|
||||
- A Claude config with one stdio server and one remote server syncs correctly to every documented MCP-capable provider.
|
||||
- No user data is deleted during sync.
|
||||
- Documentation and CLI help no longer over-promise relative to actual behavior.
|
||||
|
||||
## AI Pairing Notes
|
||||
|
||||
- Treat official provider docs as authoritative over older local notes, especially for Kiro and Qwen transport handling.
|
||||
- Have a human review any AI-generated MCP mapping code before merge because these config files may contain secrets and lossy transport assumptions are easy to miss.
|
||||
- When using an implementation agent, keep the work split by target so each provider's config contract can be tested independently.
|
||||
|
||||
## References & Research
|
||||
|
||||
### Internal References
|
||||
|
||||
- `src/commands/sync.ts:15-125`
|
||||
- `src/utils/detect-tools.ts:11-46`
|
||||
- `src/parsers/claude-home.ts:11-64`
|
||||
- `src/sync/codex.ts:7-92`
|
||||
- `src/sync/droid.ts:6-21`
|
||||
- `src/targets/windsurf.ts:13-93`
|
||||
- `src/targets/kiro.ts:5-93`
|
||||
- `src/targets/openclaw.ts:6-95`
|
||||
- `src/targets/qwen.ts:5-64`
|
||||
- `docs/solutions/adding-converter-target-providers.md:20-32`
|
||||
- `docs/solutions/adding-converter-target-providers.md:208-214`
|
||||
- `README.md:89-123`
|
||||
|
||||
### External References
|
||||
|
||||
- Factory MCP docs: https://docs.factory.ai/factory-cli/configuration/mcp
|
||||
- Factory skills docs: https://docs.factory.ai/cli/configuration/skills
|
||||
- Windsurf MCP docs: https://docs.windsurf.com/windsurf/cascade/mcp
|
||||
- Kiro MCP overview: https://kiro.dev/blog/unlock-your-development-productivity-with-kiro-and-mcp/
|
||||
- Kiro remote MCP support: https://kiro.dev/blog/introducing-remote-mcp/
|
||||
- Kiro skills announcement: https://kiro.dev/blog/custom-subagents-skills-and-enterprise-controls/
|
||||
- Qwen settings docs: https://qwenlm.github.io/qwen-code-docs/en/users/configuration/settings/
|
||||
- Qwen MCP docs: https://qwenlm.github.io/qwen-code-docs/en/users/features/mcp/
|
||||
- Qwen skills docs: https://qwenlm.github.io/qwen-code-docs/zh/users/features/skills/
|
||||
- OpenClaw setup/config docs: https://docs.openclaw.ai/start/setup
|
||||
- OpenClaw skills docs: https://docs.openclaw.ai/skills
|
||||
|
||||
## Implementation Notes for the Follow-Up `/workflows-work` Step
|
||||
|
||||
Suggested implementation order:
|
||||
|
||||
1. registry + detection cleanup
|
||||
2. codex remote MCP + droid MCP
|
||||
3. windsurf + kiro + qwen sync modules
|
||||
4. openclaw validation and implementation or explicit warning path
|
||||
5. docs + tests
|
||||
387
docs/plans/2026-03-15-001-feat-ce-ideate-skill-plan.md
Normal file
387
docs/plans/2026-03-15-001-feat-ce-ideate-skill-plan.md
Normal file
@@ -0,0 +1,387 @@
|
||||
---
|
||||
title: "feat: Add ce:ideate open-ended ideation skill"
|
||||
type: feat
|
||||
status: completed
|
||||
date: 2026-03-15
|
||||
origin: docs/brainstorms/2026-03-15-ce-ideate-skill-requirements.md
|
||||
deepened: 2026-03-16
|
||||
---
|
||||
|
||||
# feat: Add ce:ideate open-ended ideation skill
|
||||
|
||||
## Overview
|
||||
|
||||
Add a new `ce:ideate` skill to the compound-engineering plugin that performs open-ended, divergent-then-convergent idea generation for any project. The skill deeply scans the codebase, generates ~30 ideas, self-critiques and filters them, and presents the top 5-7 as a ranked list with structured analysis. It uses agent intelligence to improve the candidate pool without replacing the core prompt mechanism, writes a durable artifact to `docs/ideation/` after the survivors have been reviewed, and hands off selected ideas to `ce:brainstorm`.
|
||||
|
||||
## Problem Frame
|
||||
|
||||
The ce:* workflow pipeline has a gap at the very beginning. `ce:brainstorm` requires the user to bring an idea — it refines but doesn't generate. Users who want the AI to proactively suggest improvements must resort to ad-hoc prompting, which lacks codebase grounding, structured output, durable artifacts, and pipeline integration. (see origin: docs/brainstorms/2026-03-15-ce-ideate-skill-requirements.md)
|
||||
|
||||
## Requirements Trace
|
||||
|
||||
- R1. Standalone skill in `plugins/compound-engineering/skills/ce-ideate/`
|
||||
- R2. Optional freeform argument as focus hint (concept, path, constraint, or empty)
|
||||
- R3. Deep codebase scan via research agents before generating ideas
|
||||
- R4. Preserve the proven prompt mechanism: many ideas first, then brutal filtering, then detailed survivors
|
||||
- R5. Self-critique with explicit rejection reasoning
|
||||
- R6. Present top 5-7 with structured analysis (description, rationale, downsides, confidence 0-100%, complexity)
|
||||
- R7. Rejection summary (one-line per rejected idea)
|
||||
- R8. Durable artifact in `docs/ideation/YYYY-MM-DD-<topic>-ideation.md`
|
||||
- R9. Volume overridable via argument
|
||||
- R10. Handoff: brainstorm an idea, refine, share to Proof, or end session
|
||||
- R11. Always route to ce:brainstorm for follow-up on selected ideas
|
||||
- R12. Offer commit on session end
|
||||
- R13. Resume from existing ideation docs (30-day recency window)
|
||||
- R14. Present survivors before writing the durable artifact
|
||||
- R15. Write artifact before handoff/share/end
|
||||
- R16. Update doc in place on refine when preserving refined state
|
||||
- R17. Use agent intelligence as support for the core mechanism, not a replacement
|
||||
- R18. Use research agents for grounding; ideation/critique sub-agents are prompt-defined roles
|
||||
- R19. Pass grounding summary, focus hint, and volume target to ideation sub-agents
|
||||
- R20. Focus hints influence both generation and filtering
|
||||
- R21. Use standardized structured outputs from ideation sub-agents
|
||||
- R22. Orchestrator owns final scoring, ranking, and survivor decisions
|
||||
- R23. Use broad prompt-framing methods to encourage creative spread without over-constraining ideation
|
||||
- R24. Use the smallest useful set of sub-agents rather than a hardcoded fixed count
|
||||
- R25. Mark ideas as "explored" when brainstormed
|
||||
|
||||
## Scope Boundaries
|
||||
|
||||
- No external research (competitive analysis, similar projects) in v1 (see origin)
|
||||
- No configurable depth modes — fixed volume with argument-based override (see origin)
|
||||
- No modifications to ce:brainstorm — discovery via skill description only (see origin)
|
||||
- No deprecated `workflows:ideate` alias — the `workflows:*` prefix is deprecated
|
||||
- No `references/` split — estimated skill length ~300 lines, well under the 500-line threshold
|
||||
|
||||
## Context & Research
|
||||
|
||||
### Relevant Code and Patterns
|
||||
|
||||
- `plugins/compound-engineering/skills/ce-brainstorm/SKILL.md` — Closest sibling. Mirror: resume behavior (Phase 0.1), artifact frontmatter (date + topic), handoff options via platform question tool, document-review integration, Proof sharing
|
||||
- `plugins/compound-engineering/skills/ce-plan/SKILL.md` — Agent dispatch pattern: `Task compound-engineering:research:repo-research-analyst(context)` running in parallel. Phase 0.2 upstream document detection
|
||||
- `plugins/compound-engineering/skills/ce-work/SKILL.md` — Session completion: incremental commit pattern, staging specific files, conventional commit format
|
||||
- `plugins/compound-engineering/skills/ce-compound/SKILL.md` — Parallel research assembly: subagents return text only, orchestrator writes the single file
|
||||
- `plugins/compound-engineering/skills/document-review/SKILL.md` — Utility invocation: "Load the `document-review` skill and apply it to..." Returns "Review complete" signal
|
||||
- `plugins/compound-engineering/skills/deepen-plan/SKILL.md` — Broad parallel agent dispatch pattern
|
||||
- PR #277 (`fix: codex workflow conversion for compound-engineering`) — establishes the Codex model for canonical `ce:*` workflows: prompt wrappers for canonical entrypoints, transformed intra-workflow handoffs, and omission of deprecated `workflows:*` aliases
|
||||
|
||||
### Institutional Learnings
|
||||
|
||||
- `docs/solutions/plugin-versioning-requirements.md` — Do not bump versions or cut changelog entries in feature PRs. Do update README counts and plugin.json descriptions.
|
||||
- `docs/solutions/codex-skill-prompt-entrypoints.md` (from PR #277) — for compound-engineering workflows in Codex, prompts are the canonical user-facing entrypoints and copied skills are the reusable implementation units underneath them
|
||||
|
||||
## Key Technical Decisions
|
||||
|
||||
- **Agent dispatch for codebase scan**: Use `repo-research-analyst` + `learnings-researcher` in parallel (matches ce:plan Phase 1.1). Skip `git-history-analyzer` by default — marginal ideation value for the cost. The focus hint (R2) is passed as context to both agents.
|
||||
- **Core mechanism first, agents second**: The core design is still the user's proven prompt pattern: generate many ideas, reject aggressively, then explain only the survivors. Agent intelligence improves the candidate pool and critique quality, but does not replace this mechanism.
|
||||
- **Prompt-defined ideation and critique sub-agents**: Use prompt-shaped sub-agents with distinct framing methods for ideation and optional skeptical critique, rather than forcing reuse of existing named review agents whose purpose is different.
|
||||
- **Orchestrator-owned synthesis and scoring**: The orchestrator merges and dedupes sub-agent outputs, applies one consistent rubric, and decides final scoring/ranking. Sub-agents may emit lightweight local signals, but not authoritative final rankings.
|
||||
- **Artifact frontmatter**: `date`, `topic`, `focus` (optional). Minimal, paralleling the brainstorm `date` + `topic` pattern.
|
||||
- **Volume override via natural language**: The skill instructions tell Claude to interpret number patterns in the argument ("top 3", "100 ideas") as volume overrides. No formal parsing.
|
||||
- **Artifact timing**: Present survivors first, allow brief questions or lightweight clarification, then write/update the durable artifact before any handoff, Proof share, or session end.
|
||||
- **No `disable-model-invocation`**: The skill should be auto-loadable when users say things like "what should I improve?", "give me ideas for this project", "ideate on improvements". Following the same pattern as ce:brainstorm.
|
||||
- **Commit pattern**: Stage only `docs/ideation/<filename>`, use conventional format `docs: add ideation for <topic>`, offer but don't force.
|
||||
- **Relationship to PR #277**: `ce:ideate` must follow the same Codex workflow model as the other canonical `ce:*` workflows. Why: without #277's prompt-wrapper and handoff-rewrite model, a copied workflow skill can still point at Claude-style slash handoffs that do not exist coherently in Codex. `ce:ideate` should be introduced as another canonical `ce:*` workflow on that same surface, not as a one-off pass-through skill.
|
||||
|
||||
## Open Questions
|
||||
|
||||
### Resolved During Planning
|
||||
|
||||
- **Which agents for codebase scan?** → `repo-research-analyst` + `learnings-researcher`. Rationale: same proven pattern as ce:plan, covers both current code and institutional knowledge.
|
||||
- **Additional analysis fields per idea?** → Keep as specified in R6. "What this unlocks" bleeds into brainstorm scope. YAGNI.
|
||||
- **Volume override detection?** → Natural language interpretation. The skill instructions describe how to detect overrides. No formal parsing needed.
|
||||
- **Artifact frontmatter fields?** → `date`, `topic`, `focus` (optional). Follows brainstorm pattern.
|
||||
- **Need references/ split?** → No. Estimated ~300 lines, under the 500-line threshold.
|
||||
- **Need deprecated alias?** → No. `workflows:*` is deprecated; new skills go straight to `ce:*`.
|
||||
- **How should docs regeneration be represented in the plan?** → The checked-in tree does not currently contain the previously assumed generated files (`docs/index.html`, `docs/pages/skills.html`). Treat `/release-docs` as a repo-maintenance validation step that may update tracked generated artifacts, not as a guaranteed edit to predetermined file paths.
|
||||
- **How should skill counts be validated across artifacts?** → Do not force one unified count across every surface. The plugin manifests should reflect parser-discovered skill directories, while `plugins/compound-engineering/README.md` should preserve its human-facing taxonomy of workflow commands vs. standalone skills.
|
||||
- **What is the dependency on PR #277?** → Treat #277 as an upstream prerequisite for Codex correctness. If it merges first, `ce:ideate` should slot into its canonical `ce:*` workflow model. If it does not merge first, equivalent Codex workflow behavior must be included before `ce:ideate` is considered complete.
|
||||
- **How should agent intelligence be applied?** → Research agents are used for grounding, prompt-defined sub-agents are used to widen the candidate pool and critique it, and the orchestrator remains the final judge.
|
||||
- **Who should score the ideas?** → The orchestrator, not the ideation sub-agents and not a separate scoring sub-agent by default.
|
||||
- **When should the artifact be written?** → After the survivors are presented and reviewed enough to preserve, but always before handoff, sharing, or session end.
|
||||
|
||||
### Deferred to Implementation
|
||||
|
||||
- **Exact wording of the divergent ideation prompt section**: The plan specifies the structure and mechanisms, but the precise phrasing will be refined during implementation. This is an inherently iterative design element.
|
||||
- **Exact wording of the self-critique instructions**: Same — structure is defined, exact prose is implementation-time.
|
||||
|
||||
## Implementation Units
|
||||
|
||||
- [x] **Unit 1: Create the ce:ideate SKILL.md**
|
||||
|
||||
**Goal:** Write the complete skill definition with all phases, the ideation prompt structure, optional sub-agent support, artifact template, and handoff options.
|
||||
|
||||
**Requirements:** R1-R25 (all requirements — this is the core deliverable)
|
||||
|
||||
**Dependencies:** None
|
||||
|
||||
**Files:**
|
||||
- Create: `plugins/compound-engineering/skills/ce-ideate/SKILL.md`
|
||||
- Test (conditional): `tests/claude-parser.test.ts`, `tests/cli.test.ts`
|
||||
|
||||
**Approach:**
|
||||
|
||||
- Keep this unit primarily content-only unless implementation discovers a real parser or packaging gap. `loadClaudePlugin()` already discovers any `skills/*/SKILL.md`, and most target converters/writers already pass `plugin.skills` through as `skillDirs`.
|
||||
- Do not rely on pure pass-through for Codex. Because PR #277 gives compound-engineering `ce:*` workflows a canonical prompt-wrapper model in Codex, `ce:ideate` must be validated against that model and may require Codex-target updates if #277 is not already present.
|
||||
- Treat artifact lifecycle rules as part of the skill contract, not polish: resume detection, present-before-write, refine-in-place, and brainstorm handoff state all live inside this SKILL.md and must be internally consistent.
|
||||
- Keep the prompt sections grounded in Phase 1 findings so ideation quality does not collapse into generic product advice.
|
||||
- Keep the user's original prompt mechanism as the backbone of the workflow. Extra agent structure should strengthen that mechanism rather than replacing it.
|
||||
- When sub-agents are used, keep them prompt-defined and lightweight: shared grounding/focus/volume input, structured output, orchestrator-owned merge/dedupe/scoring.
|
||||
|
||||
The skill follows the ce:brainstorm phase structure but with fundamentally different phases:
|
||||
|
||||
```
|
||||
Phase 0: Resume and Route
|
||||
0.1 Check docs/ideation/ for recent ideation docs (R13)
|
||||
0.2 Parse argument — extract focus hint and any volume override (R2, R9)
|
||||
0.3 If no argument, proceed with fully open ideation (no blocking ask)
|
||||
|
||||
Phase 1: Codebase Scan
|
||||
1.1 Dispatch research agents in parallel (R3):
|
||||
- Task compound-engineering:research:repo-research-analyst(focus context)
|
||||
- Task compound-engineering:research:learnings-researcher(focus context)
|
||||
1.2 Consolidate scan results into a codebase understanding summary
|
||||
|
||||
Phase 2: Divergent Generation (R4, R17-R21, R23-R24)
|
||||
Core ideation instructions tell Claude to:
|
||||
- Generate ~30 ideas (or override amount) as a numbered list
|
||||
- Each idea is a one-liner at this stage
|
||||
- Push past obvious suggestions — the first 10-15 will be safe/obvious,
|
||||
the interesting ones come after
|
||||
- Ground every idea in specific codebase findings from Phase 1
|
||||
- Ideas should span multiple dimensions where justified
|
||||
- If a focus area was provided, weight toward it but don't exclude
|
||||
other strong ideas
|
||||
- Preserve the user's original many-ideas-first mechanism
|
||||
Optional sub-agent support:
|
||||
- If the platform supports it, dispatch a small useful set of ideation
|
||||
sub-agents with the same grounding summary, focus hint, and volume target
|
||||
- Give each one a distinct prompt framing method (e.g. friction, unmet
|
||||
need, inversion, assumption-breaking, leverage, extreme case)
|
||||
- Require structured idea output so the orchestrator can merge and dedupe
|
||||
- Do not use sub-agents to replace the core ideation mechanism
|
||||
|
||||
Phase 3: Self-Critique and Filter (R5, R7, R20-R22)
|
||||
Critique instructions tell Claude to:
|
||||
- Go through each idea and evaluate it critically
|
||||
- For each rejection, write a one-line reason
|
||||
- Rejection criteria: not actionable, too vague, too expensive relative
|
||||
to value, already exists, duplicates another idea, not grounded in
|
||||
actual codebase state
|
||||
- Target: keep 5-7 survivors (or override amount)
|
||||
- If more than 7 pass scrutiny, do a second pass with higher bar
|
||||
- If fewer than 5 pass, note this honestly rather than lowering the bar
|
||||
Optional critique sub-agent support:
|
||||
- Skeptical sub-agents may attack the merged list from distinct angles
|
||||
- The orchestrator synthesizes critiques and owns final scoring/ranking
|
||||
|
||||
Phase 4: Present Results (R6, R7, R14)
|
||||
- Display ranked survivors with structured analysis per idea:
|
||||
title, description (2-3 sentences), rationale, downsides,
|
||||
confidence (0-100%), estimated complexity (low/medium/high)
|
||||
- Display rejection summary: collapsed section, one-line per rejected idea
|
||||
- Allow brief questions or lightweight clarification before archival write
|
||||
|
||||
Phase 5: Write Artifact (R8, R15, R16)
|
||||
- mkdir -p docs/ideation/
|
||||
- Write the ideation doc after survivors are reviewed enough to preserve
|
||||
- Artifact includes: metadata, codebase context summary, ranked
|
||||
survivors with full analysis, rejection summary
|
||||
- Always write/update before brainstorm handoff, Proof share, or session end
|
||||
|
||||
Phase 6: Handoff (R10, R11, R12, R15-R16, R25)
|
||||
6.1 Present options via platform question tool:
|
||||
- Brainstorm an idea (pick by number → feeds to ce:brainstorm) (R11)
|
||||
- Refine (R15)
|
||||
- Share to Proof
|
||||
- End session (R12)
|
||||
6.2 Handle selection:
|
||||
- Brainstorm: update doc to mark idea as "explored" (R16),
|
||||
then invoke ce:brainstorm with the idea description
|
||||
- Refine: ask what kind of refinement, then route:
|
||||
"add more ideas" / "explore new angles" → return to Phase 2
|
||||
"re-evaluate" / "raise the bar" → return to Phase 3
|
||||
"dig deeper on idea #N" → expand that idea's analysis in place
|
||||
Update doc after each refinement when preserving the refined state (R16)
|
||||
- Share to Proof: upload ideation doc using the standard
|
||||
curl POST pattern (same as ce:brainstorm), return to options
|
||||
- End: offer to commit the ideation doc (R12), display closing summary
|
||||
```
|
||||
|
||||
Frontmatter:
|
||||
```yaml
|
||||
---
|
||||
name: ce:ideate
|
||||
description: 'Generate and critically evaluate improvement ideas for any project through deep codebase analysis and divergent-then-convergent thinking. Use when the user says "what should I improve", "give me ideas", "ideate", "surprise me with improvements", "what would you change about this project", or when they want AI-generated project improvement suggestions rather than refining their own idea.'
|
||||
argument-hint: "[optional: focus area, path, or constraint]"
|
||||
---
|
||||
```
|
||||
|
||||
Artifact template:
|
||||
```markdown
|
||||
---
|
||||
date: YYYY-MM-DD
|
||||
topic: <kebab-case-topic>
|
||||
focus: <focus area if provided, omit if open>
|
||||
---
|
||||
|
||||
# Ideation: <Topic or "Open Exploration">
|
||||
|
||||
## Codebase Context
|
||||
[Brief summary of what the scan revealed — project structure, patterns, pain points, opportunities]
|
||||
|
||||
## Ranked Ideas
|
||||
|
||||
### 1. <Idea Title>
|
||||
**Description:** [2-3 sentences]
|
||||
**Rationale:** [Why this would be a good improvement]
|
||||
**Downsides:** [Risks or costs]
|
||||
**Confidence:** [0-100%]
|
||||
**Complexity:** [Low / Medium / High]
|
||||
|
||||
### 2. <Idea Title>
|
||||
...
|
||||
|
||||
## Rejection Summary
|
||||
| # | Idea | Reason for Rejection |
|
||||
|---|------|---------------------|
|
||||
| 1 | ... | ... |
|
||||
|
||||
## Session Log
|
||||
- [Date]: Initial ideation — [N] generated, [M] survived
|
||||
```
|
||||
|
||||
**Patterns to follow:**
|
||||
- ce:brainstorm SKILL.md — phase structure, frontmatter style, argument handling, resume pattern, handoff options, Proof sharing, interaction rules
|
||||
- ce:plan SKILL.md — agent dispatch syntax (`Task compound-engineering:research:*`)
|
||||
- ce:work SKILL.md — session completion commit pattern
|
||||
- Plugin CLAUDE.md — skill compliance checklist (imperative voice, cross-platform question tool, no second person)
|
||||
|
||||
**Test scenarios:**
|
||||
- Invoke with no arguments → fully open ideation, generates ideas, presents survivors, then writes artifact when preserving results
|
||||
- Invoke with focus area (`/ce:ideate DX improvements`) → weighted ideation toward focus
|
||||
- Invoke with path (`/ce:ideate plugins/compound-engineering/skills/`) → scoped scan
|
||||
- Invoke with volume override (`/ce:ideate give me your top 3`) → adjusted volume
|
||||
- Resume: invoke when recent ideation doc exists → offers to continue or start fresh
|
||||
- Resume + refine loop: revisit an existing ideation doc, add more ideas, then re-run critique without creating a duplicate artifact
|
||||
- If sub-agents are used: each receives grounding + focus + volume context and returns structured outputs for orchestrator merge
|
||||
- If critique sub-agents are used: orchestrator remains final scorer and ranker
|
||||
- Brainstorm handoff: pick an idea → doc updated with "explored" marker, ce:brainstorm invoked
|
||||
- Refine: ask to dig deeper → doc updated in place with refined analysis
|
||||
- End session: offer commit → stages only the ideation doc, conventional message
|
||||
- Initial review checkpoint: survivors can be questioned before archival write
|
||||
- Codex install path after PR #277: `ce:ideate` is exposed as the canonical `ce:ideate` workflow entrypoint, not only as a copied raw skill
|
||||
- Codex intra-workflow handoffs: any copied `SKILL.md` references to `/ce:*` routes resolve to the canonical Codex prompt surface, and no deprecated `workflows:ideate` alias is emitted
|
||||
|
||||
**Verification:**
|
||||
- SKILL.md is under 500 lines
|
||||
- Frontmatter has `name`, `description`, `argument-hint`
|
||||
- Description includes trigger phrases for auto-discovery
|
||||
- All 25 requirements are addressed in the phase structure
|
||||
- Writing style is imperative/infinitive, no second person
|
||||
- Cross-platform question tool pattern with fallback
|
||||
- No `disable-model-invocation` (auto-loadable)
|
||||
- The repository still loads plugin skills normally because `ce:ideate` is discovered as a `skillDirs` entry
|
||||
- Codex output follows the compound-engineering workflow model from PR #277 for this new canonical `ce:*` workflow
|
||||
|
||||
---
|
||||
|
||||
- [x] **Unit 2: Update plugin metadata and documentation**
|
||||
|
||||
**Goal:** Update all locations where component counts and skill listings appear.
|
||||
|
||||
**Requirements:** R1 (skill exists in the plugin)
|
||||
|
||||
**Dependencies:** Unit 1
|
||||
|
||||
**Files:**
|
||||
- Modify: `plugins/compound-engineering/.claude-plugin/plugin.json` — update description with new skill count
|
||||
- Modify: `.claude-plugin/marketplace.json` — update plugin description with new skill count
|
||||
- Modify: `plugins/compound-engineering/README.md` — add ce:ideate to skills table/list, update count
|
||||
|
||||
**Approach:**
|
||||
- Count actual skill directories after adding ce:ideate for manifest-facing descriptions (`plugin.json`, `.claude-plugin/marketplace.json`)
|
||||
- Preserve the README's separate human-facing breakdown of `Commands` vs `Skills` instead of forcing it to equal the manifest-level skill-directory count
|
||||
- Add ce:ideate to the README skills section with a brief description in the existing table format
|
||||
- Do NOT bump version numbers (per plugin versioning requirements)
|
||||
- Do NOT add a CHANGELOG.md release entry
|
||||
|
||||
**Patterns to follow:**
|
||||
- CLAUDE.md checklist: "Updating the Compounding Engineering Plugin"
|
||||
- Existing skill entries in README.md for description format
|
||||
- `src/parsers/claude.ts` loading model: manifests and targets derive skill inventory from discovered `skills/*/SKILL.md` directories
|
||||
|
||||
**Test scenarios:**
|
||||
- Manifest descriptions reflect the post-change skill-directory count
|
||||
- README component table and skill listing stay internally consistent with the README's own taxonomy
|
||||
- JSON files remain valid
|
||||
- README skill listing includes ce:ideate
|
||||
|
||||
**Verification:**
|
||||
- `grep -o "Includes [0-9]* specialized agents" plugins/compound-engineering/.claude-plugin/plugin.json` matches actual agent count
|
||||
- Manifest-facing skill count matches the number of skill directories under `plugins/compound-engineering/skills/`
|
||||
- README counts and tables are internally consistent, even if they intentionally differ from manifest-facing skill-directory totals
|
||||
- `jq . < .claude-plugin/marketplace.json` succeeds
|
||||
- `jq . < plugins/compound-engineering/.claude-plugin/plugin.json` succeeds
|
||||
|
||||
---
|
||||
|
||||
- [x] **Unit 3: Refresh generated docs artifacts if the local docs workflow produces tracked changes**
|
||||
|
||||
**Goal:** Keep generated documentation outputs in sync without inventing source-of-truth files that are not present in the current tree.
|
||||
|
||||
**Requirements:** R1 (skill visible in docs)
|
||||
|
||||
**Dependencies:** Unit 2
|
||||
|
||||
**Files:**
|
||||
- Modify (conditional): tracked files under `docs/` updated by the local docs release workflow, if any are produced in this checkout
|
||||
|
||||
**Approach:**
|
||||
- Run the repo-maintenance docs regeneration workflow after the durable source files are updated
|
||||
- Review only the tracked artifacts it actually changes instead of assuming specific generated paths
|
||||
- If the local docs workflow produces no tracked changes in this checkout, stop without hand-editing guessed HTML files
|
||||
|
||||
**Patterns to follow:**
|
||||
- CLAUDE.md: "After ANY change to agents, commands, skills, or MCP servers, run `/release-docs`"
|
||||
|
||||
**Test scenarios:**
|
||||
- Generated docs, if present, pick up ce:ideate and updated counts from the durable sources
|
||||
- Docs regeneration does not introduce unrelated count drift across generated artifacts
|
||||
|
||||
**Verification:**
|
||||
- Any tracked generated docs diffs are mechanically consistent with the updated plugin metadata and README
|
||||
- No manual HTML edits are invented for files absent from the working tree
|
||||
|
||||
## System-Wide Impact
|
||||
|
||||
- **Interaction graph:** `ce:ideate` sits before `ce:brainstorm` and calls into `repo-research-analyst`, `learnings-researcher`, the platform question tool, optional Proof sharing, and optional local commit flow. The plan has to preserve that this is an orchestration skill spanning multiple existing workflow seams rather than a standalone document generator.
|
||||
- **Error propagation:** Resume mismatches, write-before-present failures, or refine-in-place write failures can leave the ideation artifact out of sync with what the user saw. The skill should prefer conservative routing and explicit state updates over optimistic wording.
|
||||
- **State lifecycle risks:** `docs/ideation/` becomes a new durable state surface. Topic slugging, 30-day resume matching, refinement updates, and the "explored" marker for brainstorm handoff need stable rules so repeated runs do not create duplicate or contradictory ideation records.
|
||||
- **API surface parity:** Most targets can continue to rely on copied `skillDirs`, but Codex is now a special-case workflow surface for compound-engineering because of PR #277. `ce:ideate` needs parity with the canonical `ce:*` workflow model there: explicit prompt entrypoint, rewritten intra-workflow handoffs, and no deprecated alias duplication.
|
||||
- **Integration coverage:** Unit-level reading of the SKILL.md is not enough. Verification has to cover end-to-end workflow behavior: initial ideation, artifact persistence, resume/refine loops, and handoff to `ce:brainstorm` without dropping ideation state.
|
||||
|
||||
## Risks & Dependencies
|
||||
|
||||
- **Divergent ideation quality is hard to verify at planning time**: The self-prompting instructions for Phase 2 and Phase 3 are the novel design element. Their effectiveness depends on exact wording and how well Phase 1 findings are fed back into ideation. Mitigation: verify on the real repo with open and focused prompts, then tighten the prompt structure only where groundedness or rejection quality is weak.
|
||||
- **Artifact state drift across resume/refine/handoff**: The feature depends on updating the same ideation doc repeatedly. A weak state model could duplicate docs, lose "explored" markers, or present stale survivors after refinement. Mitigation: keep one canonical ideation file per session/topic and make every refine/handoff path explicitly update that file before returning control.
|
||||
- **Count taxonomy drift across docs and manifests**: This repo already uses different count semantics across surfaces. A naive "make every number match" implementation could either break manifest descriptions or distort the README taxonomy. Mitigation: validate each artifact against its own intended counting model and document that distinction in the plan.
|
||||
- **Dependency on PR #277 for Codex workflow correctness**: `ce:ideate` is another canonical `ce:*` workflow, so its Codex install surface should not regress to the old copied-skill-only behavior. Mitigation: land #277 first or explicitly include the same Codex workflow behavior before considering this feature complete.
|
||||
- **Local docs workflow dependency**: `/release-docs` is a repo-maintenance workflow, not part of the distributed plugin. Its generated outputs may differ by environment or may not produce tracked files in the current checkout. Mitigation: treat docs regeneration as conditional maintenance verification after durable source edits, not as the primary source of truth.
|
||||
- **Skill length**: Estimated ~300 lines. If the ideation and self-critique instructions need more detail, the skill could approach the 500-line limit. Mitigation: monitor during implementation and split to `references/` only if the final content genuinely needs it.
|
||||
|
||||
## Documentation / Operational Notes
|
||||
|
||||
- README.md gets updated in Unit 2
|
||||
- Generated docs artifacts are refreshed only if the local docs workflow produces tracked changes in this checkout
|
||||
- The local `release-docs` workflow exists as a Claude slash command in this repo, but it was not directly runnable from the shell environment used for this implementation pass
|
||||
- No CHANGELOG entry for this PR (per versioning requirements)
|
||||
- No version bumps (automated release process handles this)
|
||||
|
||||
## Sources & References
|
||||
|
||||
- **Origin document:** [docs/brainstorms/2026-03-15-ce-ideate-skill-requirements.md](docs/brainstorms/2026-03-15-ce-ideate-skill-requirements.md)
|
||||
- Related code: `plugins/compound-engineering/skills/ce-brainstorm/SKILL.md`, `plugins/compound-engineering/skills/ce-plan/SKILL.md`, `plugins/compound-engineering/skills/ce-work/SKILL.md`
|
||||
- Related institutional learning: `docs/solutions/plugin-versioning-requirements.md`
|
||||
- Related PR: #277 (`fix: codex workflow conversion for compound-engineering`) — upstream Codex workflow model this plan now depends on
|
||||
- Related institutional learning: `docs/solutions/codex-skill-prompt-entrypoints.md`
|
||||
246
docs/plans/2026-03-16-001-feat-issue-grounded-ideation-plan.md
Normal file
246
docs/plans/2026-03-16-001-feat-issue-grounded-ideation-plan.md
Normal file
@@ -0,0 +1,246 @@
|
||||
---
|
||||
title: "feat: Add issue-grounded ideation mode to ce:ideate"
|
||||
type: feat
|
||||
status: active
|
||||
date: 2026-03-16
|
||||
origin: docs/brainstorms/2026-03-16-issue-grounded-ideation-requirements.md
|
||||
---
|
||||
|
||||
# feat: Add issue-grounded ideation mode to ce:ideate
|
||||
|
||||
## Overview
|
||||
|
||||
Add an issue intelligence agent and integrate it into ce:ideate so that when a user's argument indicates they want issue-tracker data as input, the skill fetches, clusters, and analyzes GitHub issues — then uses the resulting themes to drive ideation frames. The agent is also independently useful outside ce:ideate for understanding a project's issue landscape.
|
||||
|
||||
## Problem Statement / Motivation
|
||||
|
||||
ce:ideate currently grounds ideation in codebase context and past learnings only. Teams' issue trackers hold rich signal about real user pain, recurring failures, and severity patterns that ideation misses. The goal is strategic improvement ideas grounded in bug patterns ("invest in collaboration reliability") not individual bug fixes ("fix LIVE_DOC_UNAVAILABLE").
|
||||
|
||||
(See brainstorm: docs/brainstorms/2026-03-16-issue-grounded-ideation-requirements.md — R1-R9)
|
||||
|
||||
## Proposed Solution
|
||||
|
||||
Two deliverables:
|
||||
|
||||
1. **New agent**: `issue-intelligence-analyst` in `agents/research/` — fetches GitHub issues via `gh` CLI, clusters by theme, returns structured analysis. Standalone-capable.
|
||||
2. **ce:ideate modifications**: detect issue-tracker intent in arguments, dispatch the agent as a third Phase 1 scan, derive Phase 2 ideation frames from issue clusters using a hybrid strategy.
|
||||
|
||||
## Technical Approach
|
||||
|
||||
### Deliverable 1: Issue Intelligence Analyst Agent
|
||||
|
||||
**File**: `plugins/compound-engineering/agents/research/issue-intelligence-analyst.md`
|
||||
|
||||
**Frontmatter:**
|
||||
```yaml
|
||||
---
|
||||
name: issue-intelligence-analyst
|
||||
description: "Fetches and analyzes GitHub issues to surface recurring themes, pain patterns, and severity trends. Use when understanding a project's issue landscape, analyzing bug patterns for ideation, or summarizing what users are reporting."
|
||||
model: inherit
|
||||
---
|
||||
```
|
||||
|
||||
**Agent methodology (in execution order):**
|
||||
|
||||
1. **Precondition checks** — verify in order, fail fast with clear message on any failure:
|
||||
- Current directory is a git repo
|
||||
- A GitHub remote exists (prefer `upstream` over `origin` to handle fork workflows)
|
||||
- `gh` CLI is installed
|
||||
- `gh auth status` succeeds
|
||||
|
||||
2. **Fetch issues** — priority-aware, minimal fields (no bodies, no comments):
|
||||
|
||||
**Priority-aware open issue fetching:**
|
||||
- First, scan available labels to detect priority signals: `gh label list --json name --limit 100`
|
||||
- If priority/severity labels exist (e.g., `P0`, `P1`, `priority:critical`, `severity:high`, `urgent`):
|
||||
- Fetch high-priority issues first: `gh issue list --state open --label "{high-priority-labels}" --limit 50 --json number,title,labels,createdAt`
|
||||
- Backfill with remaining issues up to 100 total: `gh issue list --state open --limit 100 --json number,title,labels,createdAt` (deduplicate against already-fetched)
|
||||
- This ensures the 50 P0s in a 500-issue repo are always analyzed, not buried under 100 recent P3s
|
||||
- If no priority labels detected, fetch by recency (default `gh` sort) up to 100: `gh issue list --state open --limit 100 --json number,title,labels,createdAt`
|
||||
|
||||
**Recently closed issues:**
|
||||
- `gh issue list --state closed --limit 50 --json number,title,labels,createdAt,stateReason,closedAt` — filter client-side to last 30 days, exclude `stateReason: "not_planned"` and issues with labels matching common won't-fix patterns (`wontfix`, `won't fix`, `duplicate`, `invalid`, `by design`)
|
||||
|
||||
3. **First-pass clustering** — the core analytical step. Group issues into themes that represent **areas of systemic weakness or user pain**, not individual bugs. This is what makes the agent's output valuable.
|
||||
|
||||
**Clustering approach:**
|
||||
- Start with labels as strong clustering hints when present (e.g., `subsystem:collab` groups collaboration issues). When labels are absent or inconsistent, cluster by title similarity and inferred problem domain.
|
||||
- Cluster by **root cause or system area**, not by symptom. Example from proof repo: 25 issues mentioning `LIVE_DOC_UNAVAILABLE` and 5 mentioning `PROJECTION_STALE` are symptoms — the theme is "collaboration write path reliability." Cluster at the system level, not the error-message level.
|
||||
- Issues that span multiple themes should be noted in the primary cluster with a cross-reference, not duplicated across clusters.
|
||||
- Distinguish issue sources when relevant: bot/agent-generated issues (e.g., `agent-report` label) often have different signal quality than human-reported issues. Note the source mix per cluster — a theme with 25 agent reports and 0 human reports is different from one with 5 human reports and 2 agent reports.
|
||||
- Separate bugs from enhancement requests. Both are valid input but represent different kinds of signal (current pain vs. desired capability).
|
||||
- Aim for 3-8 themes. Fewer than 3 suggests the issues are too homogeneous or the repo has few issues. More than 8 suggests the clustering is too granular — merge related themes.
|
||||
|
||||
**What makes a good cluster:**
|
||||
- It names a systemic concern, not a specific error or ticket
|
||||
- A product or engineering leader would recognize it as "an area we need to invest in"
|
||||
- It's actionable at a strategic level (could drive an initiative, not just a patch)
|
||||
|
||||
4. **Sample body reads** — for each emerging cluster, read the full body of 2-3 representative issues (most recent or most reacted) using individual `gh issue view {number} --json body` calls. Use these to:
|
||||
- Confirm the cluster grouping is correct (titles can be misleading)
|
||||
- Understand the actual user/operator experience behind the symptoms
|
||||
- Identify severity and impact signals not captured in metadata
|
||||
- Surface any proposed solutions or workarounds already discussed
|
||||
|
||||
5. **Theme synthesis** — for each cluster, produce:
|
||||
- `theme_title`: short descriptive name
|
||||
- `description`: what the pattern is and what it signals about the system
|
||||
- `why_it_matters`: user impact, severity distribution, frequency
|
||||
- `issue_count`: number of issues in this cluster
|
||||
- `trend_direction`: increasing/stable/decreasing (compare issues opened vs closed in last 30 days within the cluster)
|
||||
- `representative_issues`: top 3 issue numbers with titles
|
||||
- `confidence`: high/medium/low based on label consistency and cluster coherence
|
||||
|
||||
6. **Return structured output** — themes ordered by issue count (descending), plus a summary line with total issues analyzed, cluster count, and date range covered.
|
||||
|
||||
**Output format (returned to caller):**
|
||||
|
||||
```markdown
|
||||
## Issue Intelligence Report
|
||||
|
||||
**Repo:** {owner/repo}
|
||||
**Analyzed:** {N} open + {M} recently closed issues ({date_range})
|
||||
**Themes identified:** {K}
|
||||
|
||||
### Theme 1: {theme_title}
|
||||
**Issues:** {count} | **Trend:** {increasing/stable/decreasing} | **Confidence:** {high/medium/low}
|
||||
|
||||
{description — what the pattern is and what it signals}
|
||||
|
||||
**Why it matters:** {user impact, severity, frequency}
|
||||
|
||||
**Representative issues:** #{num} {title}, #{num} {title}, #{num} {title}
|
||||
|
||||
### Theme 2: ...
|
||||
|
||||
### Minor / Unclustered
|
||||
{Issues that didn't fit any theme, with a brief note}
|
||||
```
|
||||
|
||||
This format is human-readable (standalone use) and structured enough for orchestrator consumption (ce:ideate use).
|
||||
|
||||
**Data source priority:**
|
||||
1. **`gh` CLI (preferred)** — most reliable, works in all terminal environments, no MCP dependency
|
||||
2. **GitHub MCP server** (fallback) — if `gh` is unavailable but a GitHub MCP server is connected, use its issue listing/reading tools instead. The clustering logic is identical; only the fetch mechanism changes.
|
||||
|
||||
If neither is available, fail gracefully per precondition checks.
|
||||
|
||||
**Token-efficient fetching:**
|
||||
|
||||
The agent runs as a sub-agent with its own context window. Every token of fetched issue data competes with the space needed for clustering reasoning. Minimize input, maximize analysis.
|
||||
|
||||
- **Metadata pass (all issues):** Fetch only the fields needed for clustering: `--json number,title,labels,createdAt,stateReason,closedAt`. Omit `body`, `comments`, `assignees`, `milestone` — these are expensive and not needed for initial grouping.
|
||||
- **Body reads (samples only):** After clusters emerge, fetch full bodies for 2-3 representative issues per cluster using individual `gh issue view {number} --json body` calls. Pick the most reacted or most recent issue in each cluster.
|
||||
- **Never fetch all bodies in bulk.** 100 issue bodies could easily consume 50k+ tokens before any analysis begins.
|
||||
|
||||
**Tool guidance** (per AGENTS.md conventions):
|
||||
- Use `gh` CLI for issue fetching (one simple command at a time, no chaining)
|
||||
- Use native file-search/glob for any repo exploration
|
||||
- Use native content-search/grep for label or pattern searches
|
||||
- Do not chain shell commands with `&&`, `||`, `;`, or pipes
|
||||
|
||||
### Deliverable 2: ce:ideate Skill Modifications
|
||||
|
||||
**File**: `plugins/compound-engineering/skills/ce-ideate/SKILL.md`
|
||||
|
||||
Four targeted modifications:
|
||||
|
||||
#### Mod 1: Phase 0.2 — Add issue-tracker intent detection
|
||||
|
||||
After the existing focus context and volume override interpretation, add a third inference:
|
||||
|
||||
- **Issue-tracker intent** — detect when the user wants issue data as input
|
||||
|
||||
The detection uses the same "reasonable interpretation rather than formal parsing" approach as the existing volume hints. Trigger on arguments whose intent is clearly about issue/bug analysis: `bugs`, `github issues`, `open issues`, `issue patterns`, `what users are reporting`, `bug reports`.
|
||||
|
||||
Do NOT trigger on arguments that merely mention bugs as a focus: `bug in auth`, `fix the login issue` — these are focus hints.
|
||||
|
||||
When combined with other dimensions (e.g., `top 3 bugs in authentication`): parse issue trigger first, volume override second, remainder is focus hint. The focus hint narrows which issues matter; the volume override controls survivor count.
|
||||
|
||||
#### Mod 2: Phase 1 — Add third parallel agent
|
||||
|
||||
Add a third numbered item to the Phase 1 parallel dispatch:
|
||||
|
||||
```
|
||||
3. **Issue intelligence** (conditional) — if issue-tracker intent was detected in Phase 0.2,
|
||||
dispatch `compound-engineering:research:issue-intelligence-analyst` with the focus hint.
|
||||
If a focus hint is present, pass it so the agent can weight its clustering.
|
||||
```
|
||||
|
||||
Update the grounding summary consolidation to include a separate **Issue Intelligence** section (distinct from codebase context) so that ideation sub-agents can distinguish between code-observed and user-reported pain points.
|
||||
|
||||
If the agent returns an error (gh not installed, no remote, auth failure), log a warning to the user ("Issue analysis unavailable: {reason}. Proceeding with standard ideation.") and continue with the existing two-agent grounding.
|
||||
|
||||
If the agent returns fewer than 5 issues total, note "Insufficient issue signal for theme analysis" and proceed with default ideation.
|
||||
|
||||
#### Mod 3: Phase 2 — Dynamic frame derivation
|
||||
|
||||
Add conditional logic before the existing frame assignment (step 8):
|
||||
|
||||
When issue-tracker intent is active and the issue intelligence agent returned themes:
|
||||
- Each theme with `confidence: high` or `confidence: medium` becomes an ideation frame. The frame prompt uses the theme title and description as the starting bias.
|
||||
- If fewer than 4 cluster-derived frames, pad with default frames selected in order: "leverage and compounding effects", "assumption-breaking or reframing", "inversion, removal, or automation of a painful step" (these complement issue-grounded themes best by pushing beyond the reported problems).
|
||||
- Cap at 6 total frames (if more than 6 themes, use the top 6 by issue count; remaining themes go into the grounding summary as "minor themes").
|
||||
|
||||
When issue-tracker intent is NOT active: existing behavior unchanged.
|
||||
|
||||
#### Mod 4: Phase 0.1 — Resume awareness
|
||||
|
||||
When checking for recent ideation documents, treat issue-grounded and non-issue ideation as distinct topics. An existing `docs/ideation/YYYY-MM-DD-open-ideation.md` should not be offered as a resume candidate when the current argument indicates issue-tracker intent, and vice versa.
|
||||
|
||||
### Files Changed
|
||||
|
||||
| File | Change |
|
||||
|------|--------|
|
||||
| `agents/research/issue-intelligence-analyst.md` | **New file** — the agent |
|
||||
| `skills/ce-ideate/SKILL.md` | **Modified** — 4 targeted modifications (Phase 0.1, 0.2, 1, 2) |
|
||||
| `.claude-plugin/plugin.json` | **Modified** — increment agent count, add agent to list, update description |
|
||||
| `../../.claude-plugin/marketplace.json` | **Modified** — update description with new agent count |
|
||||
| `README.md` | **Modified** — add agent to research agents table |
|
||||
|
||||
### Not Changed
|
||||
|
||||
- Phase 3 (adversarial filtering) — unchanged
|
||||
- Phase 4 (presentation) — unchanged, survivors already include a one-line overview
|
||||
- Phase 5 (artifact) — unchanged, the grounding summary naturally includes issue context
|
||||
- Phase 6 (refine/handoff) — unchanged
|
||||
- No other agents modified
|
||||
- No new skills
|
||||
|
||||
## Acceptance Criteria
|
||||
|
||||
- [ ] New agent file exists at `agents/research/issue-intelligence-analyst.md` with correct frontmatter
|
||||
- [ ] Agent handles precondition failures gracefully (no gh, no remote, no auth) with clear messages
|
||||
- [ ] Agent handles fork workflows (prefers upstream remote over origin)
|
||||
- [ ] Agent uses priority-aware fetching (scans for priority/severity labels, fetches high-priority first)
|
||||
- [ ] Agent caps fetching at 100 open + 50 recently closed issues
|
||||
- [ ] Agent falls back to GitHub MCP when `gh` CLI is unavailable but MCP is connected
|
||||
- [ ] Agent clusters issues into themes, not individual bug reports
|
||||
- [ ] Agent reads 2-3 sample bodies per cluster for enrichment
|
||||
- [ ] Agent output includes theme title, description, why_it_matters, issue_count, trend, representative issues, confidence
|
||||
- [ ] Agent is independently useful when dispatched directly (not just as ce:ideate sub-agent)
|
||||
- [ ] ce:ideate detects issue-tracker intent from arguments like `bugs`, `github issues`
|
||||
- [ ] ce:ideate does NOT trigger issue mode on focus hints like `bug in auth`
|
||||
- [ ] ce:ideate dispatches issue intelligence agent as third parallel Phase 1 scan when triggered
|
||||
- [ ] ce:ideate falls back to default ideation with warning when agent fails
|
||||
- [ ] ce:ideate derives ideation frames from issue clusters (hybrid: clusters + default padding)
|
||||
- [ ] ce:ideate caps at 6 frames, padding with defaults when < 4 clusters
|
||||
- [ ] Running `/ce:ideate bugs` on proof repo produces clustered themes from 25+ LIVE_DOC_UNAVAILABLE variants, not 25 separate ideas
|
||||
- [ ] Surviving ideas are strategic improvements, not individual bug fixes
|
||||
- [ ] plugin.json, marketplace.json, README.md updated with correct counts
|
||||
|
||||
## Dependencies & Risks
|
||||
|
||||
- **`gh` CLI dependency**: The agent requires `gh` installed and authenticated. Mitigated by graceful fallback to standard ideation.
|
||||
- **Issue volume**: Repos with thousands of issues could produce noisy clusters. Mitigated by fetch cap (100 open + 50 closed) and frame cap (6 max).
|
||||
- **Label quality variance**: Repos without structured labels rely on title/body clustering, which may produce lower-confidence themes. Mitigated by the confidence field and sample body reads.
|
||||
- **Context window**: Fetching 150 issues + reading 15-20 bodies could consume significant tokens in the agent's context. Mitigated by metadata-only initial fetch and sample-only body reads.
|
||||
- **Priority label detection**: No standard naming convention. Mitigated by scanning available labels and matching common patterns (P0/P1, priority:*, severity:*, urgent, critical). When no priority labels exist, falls back to recency-based fetching.
|
||||
|
||||
## Sources & References
|
||||
|
||||
- **Origin brainstorm:** [docs/brainstorms/2026-03-16-issue-grounded-ideation-requirements.md](docs/brainstorms/2026-03-16-issue-grounded-ideation-requirements.md) — Key decisions: pattern-first ideation, hybrid frame strategy, flexible argument detection, additive to Phase 1, standalone agent
|
||||
- **Exemplar agent:** `plugins/compound-engineering/agents/research/repo-research-analyst.md` — agent structure pattern
|
||||
- **ce:ideate skill:** `plugins/compound-engineering/skills/ce-ideate/SKILL.md` — integration target
|
||||
- **Institutional learning:** `docs/solutions/skill-design/compound-refresh-skill-improvements.md` — impact clustering pattern, platform-agnostic tool references, evidence-first interaction
|
||||
- **Real-world test repo:** `EveryInc/proof` (555 issues, 25+ LIVE_DOC_UNAVAILABLE duplicates, structured labels)
|
||||
@@ -0,0 +1,605 @@
|
||||
---
|
||||
title: "feat: Migrate repo releases to manual release-please with centralized changelog"
|
||||
type: feat
|
||||
status: active
|
||||
date: 2026-03-17
|
||||
origin: docs/brainstorms/2026-03-17-release-automation-requirements.md
|
||||
---
|
||||
|
||||
# feat: Migrate repo releases to manual release-please with centralized changelog
|
||||
|
||||
## Overview
|
||||
|
||||
Replace the current single-line `semantic-release` flow and maintainer-local `release-docs` workflow with a repo-owned release system built around `release-please`, a single accumulating release PR, explicit component version ownership, release automation-owned metadata/count updates, and a centralized root `CHANGELOG.md`. The new model keeps release timing manual by making merge of the generated release PR the release action while allowing dry-run previews and automatic release PR maintenance as new merges land on `main`.
|
||||
|
||||
## Problem Frame
|
||||
|
||||
The current repo mixes one automated root CLI release line with manual plugin release conventions and stale docs/tooling. `publish.yml` publishes on every push to `main`, `.releaserc.json` only understands the root package, `release-docs` still encodes outdated repo structure, and plugin-level version/changelog ownership is inconsistent. The result is drift across root changelog history, plugin manifests, computed counts, and contributor guidance. The origin requirements define a different target: manual release timing, one release PR for the whole repo, independent component versions, no bumps for untouched plugins, centralized changelog ownership, and CI-owned release authority. (see origin: docs/brainstorms/2026-03-17-release-automation-requirements.md)
|
||||
|
||||
## Requirements Trace
|
||||
|
||||
- R1. Manual release; no publish on every merge to `main`
|
||||
- R2. Batched releasable changes may accumulate on `main`
|
||||
- R3. One release PR for the whole repo that auto-accumulates releasable merges
|
||||
- R4. Independent version bumps for `cli`, `compound-engineering`, `coding-tutor`, and `marketplace`
|
||||
- R5. Untouched components do not bump
|
||||
- R6. Root `CHANGELOG.md` remains canonical
|
||||
- R7. Root changelog uses top-level component-version entries
|
||||
- R8. Existing changelog history is preserved
|
||||
- R9. `plugins/compound-engineering/CHANGELOG.md` is no longer canonical
|
||||
- R10. Retire `release-docs` as release authority
|
||||
- R11. Replace `release-docs` with narrow scripts
|
||||
- R12. Release automation owns versions, counts, and release metadata
|
||||
- R13. Support dry run with no side effects
|
||||
- R14. Dry run summarizes proposed component bumps, changelog entries, and blockers
|
||||
- R15. Marketplace version bumps only for marketplace-level changes
|
||||
- R16. Plugin version changes do not imply marketplace version bumps
|
||||
- R17. Plugin-only content changes do not force CLI version bumps
|
||||
- R18. Preserve compatibility with current install behavior where the npm CLI fetches plugin content from GitHub at runtime
|
||||
- R19. Release flow is triggerable through CI by maintainers or AI agents
|
||||
- R20. The model must scale to additional plugins
|
||||
- R21. Conventional release intent signals remain required, but component scopes in titles remain optional
|
||||
- R22. Component ownership is inferred primarily from changed files, not title scopes alone
|
||||
- R23. The repo enforces parseable conventional PR or merge titles without requiring component scope on every change
|
||||
- R24. Manual CI release supports explicit bump overrides for exceptional cases without fake commits
|
||||
- R25. Bump overrides are per-component rather than repo-wide only
|
||||
- R26. Dry run shows inferred bump and applied override clearly
|
||||
|
||||
## Scope Boundaries
|
||||
|
||||
- No change to how Claude Code consumes marketplace/plugin version fields
|
||||
- No end-user auto-update discovery flow for non-Claude harnesses in v1
|
||||
- No per-plugin canonical changelog model
|
||||
- No fully automatic timed release cadence in v1
|
||||
|
||||
## Context & Research
|
||||
|
||||
### Relevant Code and Patterns
|
||||
|
||||
- `.github/workflows/publish.yml` currently runs `npx semantic-release` on every push to `main`; this is the behavior being retired.
|
||||
- `.releaserc.json` is the current single-line release configuration and only writes `CHANGELOG.md` and `package.json`.
|
||||
- `package.json` already exposes repo-maintenance scripts and is the natural place to add release preview/validation script entrypoints.
|
||||
- `src/commands/install.ts` resolves named plugin installs by cloning the GitHub repo and reading `plugins/<name>` at runtime; this means plugin content releases can remain independent from npm CLI releases when CLI code is unchanged.
|
||||
- `.claude-plugin/marketplace.json`, `plugins/compound-engineering/.claude-plugin/plugin.json`, and `plugins/coding-tutor/.claude-plugin/plugin.json` are the current version-bearing metadata surfaces that need explicit ownership.
|
||||
- `.claude/commands/release-docs.md` is stale and mixes docs generation, metadata synchronization, validation, and release guidance; it should be replaced rather than modernized in place.
|
||||
- Existing planning docs in `docs/plans/` use one file per plan, frontmatter with `origin`, and dependency-ordered implementation units with explicit file paths; this plan follows that pattern.
|
||||
|
||||
### Institutional Learnings
|
||||
|
||||
- `docs/solutions/plugin-versioning-requirements.md` already encodes an important constraint: version bumps and changelog entries should be release-owned, not added in routine feature PRs. The migration should preserve that principle while moving the authority into CI.
|
||||
|
||||
### External References
|
||||
|
||||
- `release-please` release PR model supports maintaining a standing release PR that updates as more work lands on the default branch.
|
||||
- `release-please` manifest mode supports multi-component repos and per-component extra file updates, which is a strong fit for plugin manifests and marketplace metadata.
|
||||
- GitHub Actions `workflow_dispatch` provides a stable manual trigger surface for dry-run preview workflows.
|
||||
|
||||
## Key Technical Decisions
|
||||
|
||||
- **Use `release-please` for version planning and release PR lifecycle**: The repo needs one accumulating release PR with multiple independently versioned components; that is closer to `release-please`'s native model than to `semantic-release`.
|
||||
- **Keep one centralized root changelog**: The root `CHANGELOG.md` remains the canonical changelog. Release automation must render component-labeled entries into that one file rather than splitting canonical history across plugin-local changelog files.
|
||||
- **Use top-level component-version entries in the root changelog**: Each released component version gets its own top-level entry in `CHANGELOG.md`, including the component name, version, and release date in the heading. This keeps one centralized file while preserving readable independent version history.
|
||||
- **Treat component versioning and changelog rendering as related but separate concerns**: `release-please` can own component version bumps and release PR state, but root changelog formatting may require repo-specific rendering logic to preserve a single readable canonical file.
|
||||
- **Use explicit release scripts for repo-specific logic**: Count computation, metadata sync, dry-run summaries, and root changelog shaping should live in versioned scripts rather than hidden maintainer-local command prompts.
|
||||
- **Preserve current plugin delivery assumptions**: Plugin content updates do not force CLI version bumps unless the converter/installer behavior in `src/` changes.
|
||||
- **Marketplace is catalog-scoped**: Marketplace version bumps depend on marketplace file changes such as plugin additions/removals or marketplace metadata edits, not routine plugin release version updates.
|
||||
- **Use conventional type as release intent, not mandatory component scope**: `feat`, `fix`, and explicit breaking-change markers remain important release signals, but component scope in PR or merge titles is optional and should not be required for common compound-engineering work.
|
||||
- **File ownership is authoritative for component selection**: Optional title scope can help notes and validation, but changed-file ownership rules should decide which components bump.
|
||||
- **Support manual bump overrides as an explicit escape hatch**: Inferred bumping remains the default, but the CI-driven release flow should allow per-component `patch` / `minor` / `major` overrides for exceptional cases without requiring synthetic commits on `main`.
|
||||
- **Deprecate, do not rely on, legacy changelog/docs surfaces**: `plugins/compound-engineering/CHANGELOG.md` and `release-docs` should stop being live authorities; they should be removed, frozen, or reduced to pointer guidance only after the new flow is in place.
|
||||
|
||||
## Root Changelog Format
|
||||
|
||||
The root `CHANGELOG.md` should remain the only canonical changelog and should use component-version entries rather than repo-wide release-event entries.
|
||||
|
||||
### Format Rules
|
||||
|
||||
- Each released component gets its own top-level entry.
|
||||
- Entry headings include the component name, version, and release date.
|
||||
- Entries are ordered newest-first in the single root file.
|
||||
- When multiple components release from the same merged release PR, they appear as adjacent entries with the same date.
|
||||
- Each entry contains only changes relevant to that component.
|
||||
- The file keeps a short header note explaining that it is the canonical changelog for the repo and that versions are component-scoped.
|
||||
- Historical root changelog entries remain in place; the migration adds a note and changes formatting only for new entries after cutover.
|
||||
|
||||
### Recommended Heading Shape
|
||||
|
||||
```md
|
||||
## compound-engineering v2.43.0 - 2026-04-10
|
||||
|
||||
### Features
|
||||
- ...
|
||||
|
||||
### Fixes
|
||||
- ...
|
||||
```
|
||||
|
||||
Additional examples:
|
||||
|
||||
```md
|
||||
## coding-tutor v1.2.2 - 2026-04-18
|
||||
|
||||
### Fixes
|
||||
- ...
|
||||
|
||||
## marketplace v1.3.0 - 2026-04-18
|
||||
|
||||
### Changed
|
||||
- Added `new-plugin` to the marketplace catalog.
|
||||
|
||||
## cli v2.43.1 - 2026-04-21
|
||||
|
||||
### Fixes
|
||||
- Correct OpenClaw install path handling.
|
||||
```
|
||||
|
||||
### Migration Rules
|
||||
|
||||
- Preserve all existing root changelog history as published.
|
||||
- Add a short migration note near the top stating that, starting with the cutover release, entries are recorded per component version in the root file.
|
||||
- Do not attempt to rewrite or normalize all older entries into the new structure.
|
||||
- `plugins/compound-engineering/CHANGELOG.md` should no longer receive new canonical entries after cutover.
|
||||
|
||||
## Component Release Rules
|
||||
|
||||
The release system should use explicit file-to-component ownership rules so unchanged components do not bump accidentally.
|
||||
|
||||
### Component Definitions
|
||||
|
||||
- **`cli`**: The npm-distributed `@every-env/compound-plugin` package and its release-owned root metadata.
|
||||
- **`compound-engineering`**: The plugin rooted at `plugins/compound-engineering/`.
|
||||
- **`coding-tutor`**: The plugin rooted at `plugins/coding-tutor/`.
|
||||
- **`marketplace`**: Marketplace-level metadata rooted at `.claude-plugin/` and any future repo-owned marketplace-only surfaces.
|
||||
|
||||
### File-to-Component Mapping
|
||||
|
||||
#### `cli`
|
||||
|
||||
Changes that should trigger a `cli` release:
|
||||
|
||||
- `src/**`
|
||||
- `package.json`
|
||||
- `bun.lock`
|
||||
- CLI-only tests or fixtures that validate root CLI behavior:
|
||||
- `tests/cli.test.ts`
|
||||
- other top-level tests whose subject is the CLI itself
|
||||
- Release-owned root files only when they reflect a CLI release rather than another component:
|
||||
- root `CHANGELOG.md` entry generation for the `cli` component
|
||||
|
||||
Changes that should **not** trigger `cli` by themselves:
|
||||
|
||||
- Plugin content changes under `plugins/**`
|
||||
- Marketplace metadata changes under `.claude-plugin/**`
|
||||
- Docs or brainstorm/plan documents unless the repo explicitly decides docs-only changes are releasable for the CLI
|
||||
|
||||
#### `compound-engineering`
|
||||
|
||||
Changes that should trigger a `compound-engineering` release:
|
||||
|
||||
- `plugins/compound-engineering/**`
|
||||
- Tests or fixtures whose primary purpose is validating compound-engineering content or conversion results derived from that plugin
|
||||
- Release-owned metadata updates for the compound-engineering plugin:
|
||||
- `plugins/compound-engineering/.claude-plugin/plugin.json`
|
||||
- Root `CHANGELOG.md` entry generation for the `compound-engineering` component
|
||||
|
||||
Changes that should **not** trigger `compound-engineering` by themselves:
|
||||
|
||||
- `plugins/coding-tutor/**`
|
||||
- Root CLI implementation changes in `src/**`
|
||||
- Marketplace-only metadata changes
|
||||
|
||||
#### `coding-tutor`
|
||||
|
||||
Changes that should trigger a `coding-tutor` release:
|
||||
|
||||
- `plugins/coding-tutor/**`
|
||||
- Tests or fixtures whose primary purpose is validating coding-tutor content or conversion results derived from that plugin
|
||||
- Release-owned metadata updates for the coding-tutor plugin:
|
||||
- `plugins/coding-tutor/.claude-plugin/plugin.json`
|
||||
- Root `CHANGELOG.md` entry generation for the `coding-tutor` component
|
||||
|
||||
Changes that should **not** trigger `coding-tutor` by themselves:
|
||||
|
||||
- `plugins/compound-engineering/**`
|
||||
- Root CLI implementation changes in `src/**`
|
||||
- Marketplace-only metadata changes
|
||||
|
||||
#### `marketplace`
|
||||
|
||||
Changes that should trigger a `marketplace` release:
|
||||
|
||||
- `.claude-plugin/marketplace.json`
|
||||
- Future marketplace-only docs or config files if the repo later introduces them
|
||||
- Adding a new plugin directory under `plugins/` when that addition is accompanied by marketplace catalog changes
|
||||
- Removing a plugin from the marketplace catalog
|
||||
- Marketplace metadata changes such as owner info, catalog description, or catalog-level structure changes
|
||||
|
||||
Changes that should **not** trigger `marketplace` by themselves:
|
||||
|
||||
- Routine version bumps to existing plugin manifests
|
||||
- Plugin-only content changes under `plugins/compound-engineering/**` or `plugins/coding-tutor/**`
|
||||
- Root CLI implementation changes in `src/**`
|
||||
|
||||
### Multi-Component Rules
|
||||
|
||||
- A single merged PR may trigger multiple components when it changes files owned by each of those components.
|
||||
- A plugin content change plus a CLI behavior change should release both the plugin and `cli`.
|
||||
- Adding a new plugin should release at least the new plugin and `marketplace`; it should release `cli` only if the CLI behavior, plugin discovery logic, or install UX also changed.
|
||||
- Root `CHANGELOG.md` should not itself be used as the primary signal for component detection; it is a release output, not an input.
|
||||
- Release-owned metadata writes generated by the release flow should not recursively cause unrelated component bumps on subsequent runs.
|
||||
|
||||
### Release Intent Rules
|
||||
|
||||
- The repo should continue to require conventional release intent markers such as `feat:`, `fix:`, and explicit breaking change notation.
|
||||
- Component scopes such as `feat(coding-tutor): ...` are optional and should remain optional.
|
||||
- When a scope is present, it should be treated as advisory metadata that can improve release note grouping or mismatch detection.
|
||||
- When no scope is present, release automation should still work correctly by using changed-file ownership to determine affected components.
|
||||
- Docs-only, planning-only, or maintenance-only titles such as `docs:` or `chore:` should remain parseable even when they do not imply a releasable component bump.
|
||||
|
||||
### Manual Override Rules
|
||||
|
||||
- Automatic bump inference remains the default for all components.
|
||||
- The manual CI workflow should support override values of at least `patch`, `minor`, and `major`.
|
||||
- Overrides should be selectable per component rather than only as one repo-wide override.
|
||||
- Overrides should be treated as exceptional operational controls, not the normal release path.
|
||||
- When an override is present, release output should show both:
|
||||
- inferred bump
|
||||
- override-applied bump
|
||||
- Overrides should affect the prepared release state without requiring maintainers to add fake commits to `main`.
|
||||
|
||||
### Ambiguity Resolution Rules
|
||||
|
||||
- If a file exists primarily to support one plugin's content or fixtures, map it to that plugin rather than to `cli`.
|
||||
- If a shared utility in `src/` changes behavior for all installs/conversions, treat it as a `cli` change even if the immediate motivation came from one plugin.
|
||||
- If a change only updates docs, brainstorms, plans, or repo instructions, default to no release unless the repo intentionally adds docs-only release semantics later.
|
||||
- When a new plugin is introduced in the future, add it as its own explicit component rather than folding it into `marketplace` or `cli`.
|
||||
|
||||
## Release Workflow Behavior
|
||||
|
||||
The release flow should have three distinct modes that share the same component-detection and metadata-rendering logic.
|
||||
|
||||
### Release PR Maintenance
|
||||
|
||||
- Runs automatically on pushes to `main`.
|
||||
- Creates one release PR for the repo if none exists.
|
||||
- Updates the existing open release PR when additional releasable changes land on `main`.
|
||||
- Includes only components selected by release-intent parsing plus file ownership rules.
|
||||
- Updates release-owned files only on the release PR branch, not directly on `main`.
|
||||
- Never publishes npm, creates final GitHub releases, or tags versions as part of this maintenance step.
|
||||
|
||||
The maintained release PR should make these outputs visible:
|
||||
- component version bumps
|
||||
- draft root changelog entries
|
||||
- release-owned metadata changes such as plugin version fields and computed counts
|
||||
|
||||
### Manual Dry Run
|
||||
|
||||
- Runs only through `workflow_dispatch`.
|
||||
- Computes the same release result the current open release PR would contain, or would create if none exists.
|
||||
- Produces a human-readable summary in workflow output and optionally an artifact.
|
||||
- Validates component ownership, conventional release intent, metadata sync, count updates, and root changelog rendering.
|
||||
- Does not push commits, create or update branches, merge PRs, publish packages, create tags, or create GitHub releases.
|
||||
|
||||
The dry-run summary should include:
|
||||
- detected releasable components
|
||||
- current version -> proposed version for each component
|
||||
- draft root changelog entries
|
||||
- metadata files that would change
|
||||
- blocking validation failures and non-blocking warnings
|
||||
|
||||
### Actual Release Execution
|
||||
|
||||
- Happens only when the generated release PR is intentionally merged.
|
||||
- The merge writes the release-owned version and changelog changes into `main`.
|
||||
- Post-merge release automation then performs publish steps only for components included in that merged release.
|
||||
- npm publish runs only when the `cli` component is part of the merged release.
|
||||
- Non-CLI component releases still update canonical version surfaces and release notes even when no npm publish occurs.
|
||||
|
||||
### Safety Rules
|
||||
|
||||
- Ordinary feature merges to `main` must never publish by themselves.
|
||||
- Dry run must remain side-effect free.
|
||||
- Release PR maintenance, dry run, and post-merge release must use the same underlying release-state computation.
|
||||
- Release-generated version and metadata writes must not recursively trigger a follow-up release that contains only its own generated churn.
|
||||
- The release PR merge remains the auditable manual boundary; do not replace it with direct-to-main release commits from a manual workflow.
|
||||
|
||||
## Open Questions
|
||||
|
||||
### Resolved During Planning
|
||||
|
||||
- **Should release timing remain manual?** Yes. The release PR may be maintained automatically, but release happens only when the generated release PR is intentionally merged.
|
||||
- **Should the release PR update automatically as more merges land on `main`?** Yes. This is a core batching behavior and should remain automatic.
|
||||
- **Should release preview be distinct from release execution?** Yes. Dry run should be a side-effect-free manual workflow that previews the same release state without mutating branches or publishing anything.
|
||||
- **Should root changelog history stay centralized?** Yes. The root `CHANGELOG.md` remains canonical to avoid fragmented history.
|
||||
- **What changelog structure best fits the centralized model?** Top-level component-version entries in the root changelog are the preferred format. This keeps the file centralized while making independent version history readable.
|
||||
- **What should drive component bumps?** Explicit file-to-component ownership rules. `src/**` drives `cli`, each `plugins/<name>/**` tree drives its own plugin, and `.claude-plugin/marketplace.json` drives `marketplace`.
|
||||
- **How strict should conventional formatting be?** Conventional type should be required strongly enough for release tooling and release-note generation, but component scope should remain optional to match the repo's work style.
|
||||
- **Should exceptional manual bumping be supported?** Yes. The release workflow should expose per-component patch/minor/major override controls rather than forcing synthetic commits to manipulate inferred versions.
|
||||
- **Should marketplace version bump when only a listed plugin version changes?** No. Marketplace bumps are reserved for marketplace-level changes.
|
||||
- **Should `release-docs` remain part of release authority?** No. It should be retired and replaced with narrow scripts.
|
||||
|
||||
### Deferred to Implementation
|
||||
|
||||
- What exact combination of `release-please` config and custom post-processing yields the chosen root changelog output without fighting the tool too hard?
|
||||
- Should conventional-format enforcement happen on PR titles, squash-merge titles, commit messages, or a combination of them?
|
||||
- Should `plugins/compound-engineering/CHANGELOG.md` be deleted outright or replaced with a short pointer note after the migration is stable?
|
||||
- Should release preview be implemented by invoking `release-please` in dry-run mode directly, or by a repo-owned script that computes the same summary from component rules and current git state?
|
||||
- Should final post-merge release execution live in a dedicated publish workflow keyed off merged release PR state, or remain in a renamed/adapted version of the current `publish.yml`?
|
||||
- Should override inputs be encoded directly into release workflow inputs only, or also persisted into the generated release PR body for auditability?
|
||||
|
||||
## Implementation Units
|
||||
|
||||
- [x] **Unit 1: Define the new release component model and config scaffolding**
|
||||
|
||||
**Goal:** Replace the single-line semantic-release configuration with release-please-oriented repo configuration that expresses the four release components and their version surfaces.
|
||||
|
||||
**Requirements:** R1, R3, R4, R5, R15, R16, R17, R20
|
||||
|
||||
**Dependencies:** None
|
||||
|
||||
**Files:**
|
||||
- Create: `.release-please-config.json`
|
||||
- Create: `.release-please-manifest.json`
|
||||
- Modify: `package.json`
|
||||
- Modify: `.github/workflows/publish.yml`
|
||||
- Delete or freeze: `.releaserc.json`
|
||||
|
||||
**Approach:**
|
||||
- Define components for `cli`, `compound-engineering`, `coding-tutor`, and `marketplace`.
|
||||
- Use manifest configuration so version lines are independent and untouched components do not bump.
|
||||
- Rework the existing publish workflow so it no longer releases on every push to `main` and instead supports the release-please-driven model.
|
||||
- Add package scripts for release preview, metadata sync, and validation so CI can call stable entrypoints instead of embedding release logic inline.
|
||||
- Define the repo's release-intent contract: conventional type required, breaking changes explicit, component scope optional, file ownership authoritative.
|
||||
- Define the override contract: per-component `auto | patch | minor | major`, with `auto` as the default.
|
||||
|
||||
**Patterns to follow:**
|
||||
- Existing repo-level config files at the root (`package.json`, `.releaserc.json`, `.github/workflows/*.yml`)
|
||||
- Current release ownership documented in `docs/solutions/plugin-versioning-requirements.md`
|
||||
|
||||
**Test scenarios:**
|
||||
- A plugin-only change maps to that plugin component without implying CLI or marketplace bump.
|
||||
- A marketplace metadata/catalog change maps to marketplace only.
|
||||
- A `src/` CLI behavior change maps to the CLI component.
|
||||
- A combined change yields multiple component updates inside one release PR.
|
||||
- A title like `fix: adjust ce:plan-beta wording` remains valid without component scope and still produces the right component mapping from files.
|
||||
- A manual override can promote an inferred patch bump for one component to minor without affecting unrelated components.
|
||||
|
||||
**Verification:**
|
||||
- The repo contains a single authoritative release configuration model for all versioned components.
|
||||
- The old automatic-on-push semantic-release path is removed or inert.
|
||||
- Package scripts exist for preview/sync/validate entrypoints.
|
||||
- Release intent rules are documented without forcing repetitive component scoping on routine CE work.
|
||||
|
||||
- [x] **Unit 2: Build repo-owned release scripts for metadata sync, counts, and preview**
|
||||
|
||||
**Goal:** Replace `release-docs` and ad-hoc release bookkeeping with explicit scripts that compute release-owned metadata updates and produce dry-run summaries.
|
||||
|
||||
**Requirements:** R10, R11, R12, R13, R14, R18, R19
|
||||
|
||||
**Dependencies:** Unit 1
|
||||
|
||||
**Files:**
|
||||
- Create: `scripts/release/sync-metadata.ts`
|
||||
- Create: `scripts/release/render-root-changelog.ts`
|
||||
- Create: `scripts/release/preview.ts`
|
||||
- Create: `scripts/release/validate.ts`
|
||||
- Modify: `package.json`
|
||||
|
||||
**Approach:**
|
||||
- `sync-metadata.ts` should own count calculation and synchronized writes to release-owned metadata fields such as manifest descriptions and version mirrors.
|
||||
- `render-root-changelog.ts` should generate the centralized root changelog entries in the agreed component-version format.
|
||||
- `preview.ts` should summarize proposed component bumps, generated changelog entries, affected files, and validation blockers without mutating the repo or publishing anything.
|
||||
- `validate.ts` should provide a stable CI check for component counts, manifest consistency, and changelog formatting expectations.
|
||||
- `preview.ts` should accept optional per-component overrides and display both inferred and effective bump levels in its summary output.
|
||||
|
||||
**Patterns to follow:**
|
||||
- TypeScript/Bun scripting already used elsewhere in the repo
|
||||
- Root package scripts as stable repo entrypoints
|
||||
|
||||
**Test scenarios:**
|
||||
- Count calculation updates plugin descriptions correctly when agents/skills change.
|
||||
- Preview output includes only changed components.
|
||||
- Preview mode performs no file writes.
|
||||
- Validation fails when manifest counts or version ownership rules drift.
|
||||
- Root changelog renderer produces component-version entries with stable ordering and headings.
|
||||
- Preview output clearly distinguishes inferred bump from override-applied bump when an override is used.
|
||||
|
||||
**Verification:**
|
||||
- `release-docs` responsibilities are covered by explicit scripts.
|
||||
- Dry run can run in CI without side effects.
|
||||
- Metadata/count drift can be detected deterministically before release.
|
||||
|
||||
- [x] **Unit 3: Wire release PR maintenance and manual release execution in CI**
|
||||
|
||||
**Goal:** Establish one standing release PR for the repo that updates automatically as new releasable work lands, while keeping the actual release action manual.
|
||||
|
||||
**Requirements:** R1, R2, R3, R13, R14, R19
|
||||
|
||||
**Dependencies:** Units 1-2
|
||||
|
||||
**Files:**
|
||||
- Create: `.github/workflows/release-pr.yml`
|
||||
- Create: `.github/workflows/release-preview.yml`
|
||||
- Modify: `.github/workflows/ci.yml`
|
||||
- Modify: `.github/workflows/publish.yml`
|
||||
|
||||
**Approach:**
|
||||
- `release-pr.yml` should run on push to `main` and maintain the standing release PR for the whole repo.
|
||||
- The actual release event should remain merge of that generated release PR; no automatic publish should happen on ordinary merges to `main`.
|
||||
- `release-preview.yml` should use `workflow_dispatch` with explicit dry-run inputs and publish a human-readable summary to workflow logs and/or artifacts.
|
||||
- Decide whether npm publish remains in `publish.yml` or moves into the release-please-driven workflow, but ensure it runs only when the CLI component is actually releasing.
|
||||
- Keep normal `ci.yml` focused on verification, not publishing.
|
||||
- Add lightweight validation for release-intent formatting on PR or merge titles, without requiring component scopes.
|
||||
- Ensure release PR maintenance, dry run, and post-merge publish all call the same underlying release-state computation so they cannot drift.
|
||||
- Add workflow inputs for per-component bump overrides and ensure they can shape the prepared release state when explicitly invoked by a maintainer or AI agent.
|
||||
|
||||
**Patterns to follow:**
|
||||
- Existing GitHub workflow layout in `.github/workflows/`
|
||||
- Current manual `workflow_dispatch` presence in `publish.yml`
|
||||
|
||||
**Test scenarios:**
|
||||
- A normal merge to `main` updates or creates the release PR but does not publish.
|
||||
- A manual dry-run workflow produces a summary with no tags, commits, or publishes.
|
||||
- Merging the release PR results in release creation for changed components only.
|
||||
- A release that excludes CLI does not attempt npm publish.
|
||||
- A PR titled `feat: add new plan-beta handoff guidance` passes validation without a component scope.
|
||||
- A PR titled with an explicit contradictory scope can be surfaced as a warning or failure if file ownership clearly disagrees.
|
||||
- A second releasable merge to `main` updates the existing open release PR instead of creating a competing release PR.
|
||||
- A dry run executed while a release PR is open reports the same proposed component set and versions as the PR contents.
|
||||
- Merging a release PR does not immediately create a follow-up release PR containing only release-generated metadata churn.
|
||||
- A manual workflow can override one component to `major` while leaving other components on inferred `auto`.
|
||||
|
||||
**Verification:**
|
||||
- Maintainers can inspect the current release PR to see the pending release batch.
|
||||
- Dry-run and actual-release paths are distinct and safe.
|
||||
- The release system is triggerable through CI without local maintainer-only tooling.
|
||||
- The same proposed release state is visible consistently across release PR maintenance, dry run, and post-merge release execution.
|
||||
- Exceptional release overrides are possible without synthetic commits on `main`.
|
||||
|
||||
- [x] **Unit 4: Centralize changelog ownership and retire plugin-local canonical release history**
|
||||
|
||||
**Goal:** Make the root changelog the only canonical changelog while preserving history and preventing future fragmentation.
|
||||
|
||||
**Requirements:** R6, R7, R8, R9
|
||||
|
||||
**Dependencies:** Units 1-3
|
||||
|
||||
**Files:**
|
||||
- Modify: `CHANGELOG.md`
|
||||
- Modify or replace: `plugins/compound-engineering/CHANGELOG.md`
|
||||
- Optionally create: `plugins/coding-tutor/CHANGELOG.md` only if needed as a non-canonical pointer or future placeholder
|
||||
|
||||
**Approach:**
|
||||
- Add a migration note near the top of the root changelog clarifying that it is the canonical changelog for the repo and future releases.
|
||||
- Render future canonical entries into the root file as top-level component-version entries using the agreed heading shape.
|
||||
- Stop writing future canonical entries into `plugins/compound-engineering/CHANGELOG.md`.
|
||||
- Replace the plugin-local changelog with either a short pointer note or a frozen historical file, depending on the least confusing path discovered during implementation.
|
||||
- Keep existing root changelog entries intact; do not attempt to rewrite historical releases into a new structure retroactively.
|
||||
|
||||
**Patterns to follow:**
|
||||
- Existing Keep a Changelog-style root file
|
||||
- Brainstorm decision favoring centralized history over fragmented per-plugin changelogs
|
||||
|
||||
**Test scenarios:**
|
||||
- Historical root changelog entries remain intact after migration.
|
||||
- New generated entries appear in the root changelog in the intended component-version format.
|
||||
- Multiple components released on the same day appear as separate adjacent entries rather than being merged into one release-event block.
|
||||
- Component-specific notes do not leak unrelated changes into the wrong entry.
|
||||
- Plugin-local CE changelog no longer acts as a live release target.
|
||||
|
||||
**Verification:**
|
||||
- A maintainer reading the repo can identify one canonical changelog without ambiguity.
|
||||
- No history is lost or silently rewritten.
|
||||
|
||||
- [x] **Unit 5: Remove legacy release guidance and replace it with the new authority model**
|
||||
|
||||
**Goal:** Update repo instructions and docs so contributors follow the new release system rather than obsolete semantic-release or `release-docs` guidance.
|
||||
|
||||
**Requirements:** R10, R11, R12, R19, R20
|
||||
|
||||
**Dependencies:** Units 1-4
|
||||
|
||||
**Files:**
|
||||
- Modify: `AGENTS.md`
|
||||
- Modify: `CLAUDE.md`
|
||||
- Modify: `plugins/compound-engineering/AGENTS.md`
|
||||
- Modify: `docs/solutions/plugin-versioning-requirements.md`
|
||||
- Delete: `.claude/commands/release-docs.md` or replace with a deprecation stub
|
||||
|
||||
**Approach:**
|
||||
- Update all contributor-facing docs so they describe release PR maintenance, manual release merge, centralized root changelog ownership, and the new scripts for sync/preview/validate.
|
||||
- Remove references that tell contributors to run `release-docs` or to rely on stale docs-generation assumptions.
|
||||
- Keep the contributor rule that release-owned metadata should not be hand-bumped in ordinary PRs, but point that rule at release automation rather than a local maintainer slash command.
|
||||
- Document the release-intent policy explicitly: conventional type required, component scope optional, breaking changes explicit.
|
||||
|
||||
**Patterns to follow:**
|
||||
- Existing contributor guidance files already used as authoritative workflow docs
|
||||
|
||||
**Test scenarios:**
|
||||
- No user-facing doc still points to `release-docs` as a required release workflow.
|
||||
- No contributor guidance still claims plugin-local changelog authority for CE.
|
||||
- Release ownership guidance is consistent across root and plugin-level instruction files.
|
||||
|
||||
**Verification:**
|
||||
- A new maintainer can understand the release process from docs alone without hidden local workflows.
|
||||
- Docs no longer encode obsolete repo structure or stale release surfaces.
|
||||
|
||||
- [x] **Unit 6: Add automated coverage for component detection, metadata sync, and release preview**
|
||||
|
||||
**Goal:** Protect the new release model against regression by testing the component rules, metadata updates, and preview behavior.
|
||||
|
||||
**Requirements:** R4, R5, R12, R13, R14, R15, R16, R17
|
||||
|
||||
**Dependencies:** Units 1-5
|
||||
|
||||
**Files:**
|
||||
- Create: `tests/release-metadata.test.ts`
|
||||
- Create: `tests/release-preview.test.ts`
|
||||
- Create: `tests/release-components.test.ts`
|
||||
- Modify: `package.json`
|
||||
|
||||
**Approach:**
|
||||
- Add fixture-driven tests for file-change-to-component mapping.
|
||||
- Snapshot or assert dry-run summaries for representative release cases.
|
||||
- Verify metadata sync updates only expected files and counts.
|
||||
- Cover the marketplace-specific rule so plugin-only version changes do not trigger marketplace bumps.
|
||||
- Encode ambiguity-resolution cases explicitly so future contributors can add new plugins without guessing which component should bump.
|
||||
- Add validation coverage for release-intent parsing so conventional titles remain required but optional scopes remain non-blocking when omitted.
|
||||
- Add override-path coverage so manual bump overrides remain scoped, visible, and side-effect free in preview mode.
|
||||
|
||||
**Patterns to follow:**
|
||||
- Existing top-level Bun test files under `tests/`
|
||||
- Current fixture-driven testing style used by converters and writers
|
||||
|
||||
**Test scenarios:**
|
||||
- Change only `plugins/coding-tutor/**` and confirm only `coding-tutor` bumps.
|
||||
- Change only `plugins/compound-engineering/**` and confirm only CE bumps.
|
||||
- Change only marketplace catalog metadata and confirm only marketplace bumps.
|
||||
- Change only `src/**` and confirm only CLI bumps.
|
||||
- Combined `src/**` + plugin change yields both component bumps.
|
||||
- Change docs only and confirm no component bumps by default.
|
||||
- Add a new plugin directory plus marketplace catalog entry and confirm new-plugin + marketplace bump without forcing unrelated existing plugin bumps.
|
||||
- Dry-run preview lists the same components that the component detector identifies.
|
||||
- Conventional `fix:` / `feat:` titles without scope pass validation.
|
||||
- Explicit breaking-change markers are recognized.
|
||||
- Optional scopes, when present, can be compared against file ownership without becoming mandatory.
|
||||
- Override one component in preview and confirm only that component's effective bump changes.
|
||||
- Override does not create phantom bumps for untouched components.
|
||||
|
||||
**Verification:**
|
||||
- The release model is covered by automated tests rather than only CI trial runs.
|
||||
- Future plugin additions can follow the same component-detection pattern with low risk.
|
||||
|
||||
## System-Wide Impact
|
||||
|
||||
- **Interaction graph:** Release config, CI workflows, metadata-bearing JSON files, contributor docs, and changelog generation are all coupled. The plan deliberately separates configuration, scripting, release PR maintenance, and documentation cleanup so one layer can change without obscuring another.
|
||||
- **Error propagation:** Release metadata drift should fail in preview/validation before a release PR or publish path proceeds. CI needs clear failure reporting because release mistakes affect user-facing version surfaces.
|
||||
- **State lifecycle risks:** Partial migration is risky. Running old and new release authorities simultaneously could double-write changelog entries, version fields, or publish flows. The migration should explicitly disable the old path before trusting the new one.
|
||||
- **API surface parity:** Contributor-facing workflows in `AGENTS.md`, `CLAUDE.md`, and plugin-level instructions must all describe the same release authority model or maintainers will continue using legacy local commands.
|
||||
- **Integration coverage:** Unit tests for scripts are not enough. The workflow interaction between release PR maintenance, dry-run preview, and conditional CLI publish needs at least one integration-level verification path in CI.
|
||||
|
||||
## Risks & Dependencies
|
||||
|
||||
- `release-please` may not natively express the exact root changelog shape you want; custom rendering may be required.
|
||||
- If old semantic-release and new release-please flows overlap during migration, duplicate or conflicting release writes are likely.
|
||||
- The distinction between version-bearing metadata and descriptive/count-bearing metadata must stay explicit; otherwise scripts may overwrite user-edited documentation that should remain manual.
|
||||
- Release preview quality matters. If dry run is vague or noisy, maintainers will bypass it and the manual batching goal will weaken.
|
||||
- Removing `release-docs` may expose other hidden docs/deploy assumptions, especially if GitHub Pages or docs generation still depend on stale paths.
|
||||
|
||||
## Documentation / Operational Notes
|
||||
|
||||
- Document one canonical release path: release PR maintenance on push to `main`, dry-run preview on manual dispatch, actual release on merge of the generated release PR.
|
||||
- Document one canonical changelog: root `CHANGELOG.md`.
|
||||
- Document one rule for contributors: ordinary feature PRs do not hand-bump release-owned versions or changelog entries.
|
||||
- Add a short migration note anywhere old release instructions are likely to be rediscovered, especially around `plugins/compound-engineering/CHANGELOG.md` and the removed `release-docs` command.
|
||||
- After merge, run one live GitHub Actions validation pass to confirm `release-please` tag/output wiring and conditional CLI publish behavior end to end.
|
||||
|
||||
## Sources & References
|
||||
|
||||
- **Origin document:** [docs/brainstorms/2026-03-17-release-automation-requirements.md](docs/brainstorms/2026-03-17-release-automation-requirements.md)
|
||||
- Existing release workflow: `.github/workflows/publish.yml`
|
||||
- Existing semantic-release config: `.releaserc.json`
|
||||
- Existing release-owned guidance: `docs/solutions/plugin-versioning-requirements.md`
|
||||
- Legacy repo-maintenance command to retire: `.claude/commands/release-docs.md`
|
||||
- Install behavior reference: `src/commands/install.ts`
|
||||
- External docs: `release-please` manifest and release PR documentation, GitHub Actions `workflow_dispatch`
|
||||
574
docs/plans/feature_opencode-commands-as-md-and-config-merge.md
Normal file
574
docs/plans/feature_opencode-commands-as-md-and-config-merge.md
Normal file
@@ -0,0 +1,574 @@
|
||||
# Feature: OpenCode Commands as .md Files, Config Merge, and Permissions Default Fix
|
||||
|
||||
**Type:** feature + bug fix (consolidated)
|
||||
**Date:** 2026-02-20
|
||||
**Starting point:** Branch `main` at commit `174cd4c`
|
||||
**Create feature branch:** `feature/opencode-commands-md-merge-permissions`
|
||||
**Baseline tests:** 180 pass, 0 fail (run `bun test` to confirm before starting)
|
||||
|
||||
---
|
||||
|
||||
## Context
|
||||
|
||||
### User-Facing Goal
|
||||
|
||||
When running `bunx @every-env/compound-plugin install compound-engineering --to opencode`, three problems exist:
|
||||
|
||||
1. **Commands overwrite `opencode.json`**: Plugin commands are written into the `command` key of `opencode.json`, which replaces the user's existing configuration file (the writer does `writeJson(configPath, bundle.config)` — a full overwrite). The user loses their personal settings (model, theme, provider keys, MCP servers they previously configured).
|
||||
|
||||
2. **Commands should be `.md` files, not JSON**: OpenCode supports defining commands as individual `.md` files in `~/.config/opencode/commands/`. This is additive and non-destructive — one file per command, never touches `opencode.json`.
|
||||
|
||||
3. **`--permissions broad` is the default and pollutes global config**: The `--permissions` flag defaults to `"broad"`, which writes 14 `permission: allow` entries and 14 `tools: true` entries into `opencode.json` on every install. These are global settings that affect ALL OpenCode sessions, not just plugin commands. Even `--permissions from-commands` is semantically wrong — it unions per-command `allowedTools` restrictions into a single global block, which inverts restriction semantics (a command allowing only `Read` gets merged with one allowing `Bash`, producing global `bash: allow`).
|
||||
|
||||
### Expected Behavior After This Plan
|
||||
|
||||
- Commands are written as `~/.config/opencode/commands/<name>.md` with YAML frontmatter (`description`, `model`). The `command` key is never written to `opencode.json`.
|
||||
- `opencode.json` is deep-merged (not overwritten): existing user keys survive, plugin's MCP servers are added. User values win on conflict.
|
||||
- `--permissions` defaults to `"none"` — no `permission` or `tools` entries are written to `opencode.json` unless the user explicitly passes `--permissions broad` or `--permissions from-commands`.
|
||||
|
||||
### Relevant File Paths
|
||||
|
||||
| File | Current State on `main` | What Changes |
|
||||
|---|---|---|
|
||||
| `src/types/opencode.ts` | `OpenCodeBundle` has no `commandFiles` field. Has `OpenCodeCommandConfig` type and `command` field on `OpenCodeConfig`. | Add `OpenCodeCommandFile` type. Add `commandFiles` to `OpenCodeBundle`. Remove `OpenCodeCommandConfig` type and `command` field from `OpenCodeConfig`. |
|
||||
| `src/converters/claude-to-opencode.ts` | `convertCommands()` returns `Record<string, OpenCodeCommandConfig>`. Result set on `config.command`. `applyPermissions()` writes `config.permission` and `config.tools`. | `convertCommands()` returns `OpenCodeCommandFile[]`. `config.command` is never set. No changes to `applyPermissions()` itself. |
|
||||
| `src/targets/opencode.ts` | `writeOpenCodeBundle()` does `writeJson(configPath, bundle.config)` — full overwrite. No `commandsDir`. No merge logic. | Add `commandsDir` to path resolver. Write command `.md` files with backup. Replace overwrite with `mergeOpenCodeConfig()` — read existing, deep-merge, write back. |
|
||||
| `src/commands/install.ts` | `--permissions` default is `"broad"` (line 51). | Change default to `"none"`. Update description string. |
|
||||
| `src/utils/files.ts` | Has `readJson()`, `pathExists()`, `backupFile()` already. | No changes needed — utilities already exist. |
|
||||
| `tests/converter.test.ts` | Tests reference `bundle.config.command` (lines 19, 74, 202-214, 243). Test `"maps commands, permissions, and agents"` tests `from-commands` mode. | Update all to use `bundle.commandFiles`. Rename permission-related test to clarify opt-in nature. |
|
||||
| `tests/opencode-writer.test.ts` | 4 tests, none have `commandFiles` in bundles. `"backs up existing opencode.json before overwriting"` test expects full overwrite. | Add `commandFiles: []` to all existing bundles. Rewrite backup test to test merge behavior. Add new tests for command file writing and merge. |
|
||||
| `tests/cli.test.ts` | 10 tests. None check for commands directory. | Add test for `--permissions none` default. Add test for command `.md` file existence. |
|
||||
| `AGENTS.md` | Line 10: "Keep OpenCode output at `opencode.json` and `.opencode/{agents,skills,plugins}`." | Update to document commands go to `commands/<name>.md`, `opencode.json` is deep-merged. |
|
||||
| `README.md` | Line 54: "OpenCode output is written to `~/.config/opencode` by default, with `opencode.json` at the root..." | Update to document `.md` command files, merge behavior, `--permissions` default. |
|
||||
|
||||
### Prior Context (Pre-Investigation)
|
||||
|
||||
- **No `docs/decisions/` directory on `main`**: ADRs will be created fresh during this plan.
|
||||
- **No prior plans touch the same area**: The `2026-02-08-feat-convert-local-md-settings-for-opencode-codex-plan.md` discusses path rewriting in command bodies but does not touch command output format or permissions.
|
||||
- **OpenCode docs (confirmed via context7 MCP, library `/sst/opencode`):**
|
||||
- Command `.md` frontmatter supports: `description`, `agent`, `model`. Does NOT support `permission` or `tools`. Placed in `~/.config/opencode/commands/` (global) or `.opencode/commands/` (project).
|
||||
- Agent `.md` frontmatter supports: `description`, `mode`, `model`, `temperature`, `tools`, `permission`. Placed in `~/.config/opencode/agents/` or `.opencode/agents/`.
|
||||
- `opencode.json` is the only place for: `mcp`, global `permission`, global `tools`, `model`, `provider`, `theme`, `server`, `compaction`, `watcher`, `share`.
|
||||
|
||||
### Rejected Approaches
|
||||
|
||||
**1. Map `allowedTools` to per-agent `.md` frontmatter permissions.**
|
||||
Rejected: Claude commands are not agents. There is no per-command-to-per-agent mapping. Commands don't specify which agent to run with. Even if they did, the union of multiple commands' restrictions onto a single agent's permissions loses the per-command scoping. Agent `.md` files DO support `permission` in frontmatter, but this would require creating synthetic agents just to hold permissions — misleading and fragile.
|
||||
|
||||
**2. Write permissions into command `.md` file frontmatter.**
|
||||
Rejected: OpenCode command `.md` files only support `description`, `agent`, `model` in frontmatter. There is no `permission` or `tools` key. Confirmed via context7 docs. Anything else is silently ignored.
|
||||
|
||||
**3. Keep `from-commands` as the default but fix the flattening logic.**
|
||||
Rejected: There is no correct way to flatten per-command tool restrictions into a single global permission block. Any flattening loses information and inverts semantics.
|
||||
|
||||
**4. Remove the `--permissions` flag entirely.**
|
||||
Rejected: Some users may want to write permissions to `opencode.json` as a convenience. Keeping the flag with a changed default preserves optionality.
|
||||
|
||||
**5. Write commands as both `.md` files AND in `opencode.json` `command` block.**
|
||||
Rejected: Redundant and defeats the purpose of avoiding `opencode.json` pollution. `.md` files are the sole output format.
|
||||
|
||||
---
|
||||
|
||||
## Decision Record
|
||||
|
||||
### Decision 1: Commands emitted as individual `.md` files, never in `opencode.json`
|
||||
|
||||
- **Decision:** `convertCommands()` returns `OpenCodeCommandFile[]` (one `.md` file per command with YAML frontmatter). The `command` key is never set on `OpenCodeConfig`. The writer creates `<commandsDir>/<name>.md` for each file.
|
||||
- **Context:** OpenCode supports two equivalent formats for commands — JSON in config and `.md` files. The `.md` format is additive (new files) rather than destructive (rewriting JSON). This is consistent with how agents and skills are already handled as `.md` files.
|
||||
- **Alternatives rejected:** JSON-only (destructive), both formats (redundant). See Rejected Approaches above.
|
||||
- **Assumptions:** OpenCode resolves commands from the `commands/` directory at runtime. Confirmed via docs.
|
||||
- **Reversal trigger:** If OpenCode deprecates `.md` command files or the format changes incompatibly.
|
||||
|
||||
### Decision 2: `opencode.json` deep-merged, not overwritten
|
||||
|
||||
- **Decision:** `writeOpenCodeBundle()` reads the existing `opencode.json` (if present), deep-merges plugin-provided keys (MCP servers, and optionally permission/tools if `--permissions` is not `none`) without overwriting user-set values, and writes the merged result. User keys always win on conflict.
|
||||
- **Context:** Users have personal configuration in `opencode.json` (API keys, model preferences, themes, existing MCP servers). The current full-overwrite destroys all of this.
|
||||
- **Alternatives rejected:** Skip writing `opencode.json` entirely — rejected because MCP servers must be written there (no `.md` alternative exists for MCP).
|
||||
- **Assumptions:** `readJson()` and `pathExists()` already exist in `src/utils/files.ts`. Malformed JSON in existing file should warn and fall back to plugin-only config (do not crash, do not destroy).
|
||||
- **Reversal trigger:** If OpenCode adds a separate mechanism for plugin MCP server registration that doesn't involve `opencode.json`.
|
||||
|
||||
### Decision 3: `--permissions` default changed from `"broad"` to `"none"`
|
||||
|
||||
- **Decision:** The `--permissions` CLI flag default changes from `"broad"` to `"none"`. No `permission` or `tools` keys are written to `opencode.json` unless the user explicitly opts in.
|
||||
- **Context:** `"broad"` silently writes 14 global tool permissions. `"from-commands"` has a semantic inversion bug (unions per-command restrictions into global allows). Both are destructive to user config. `applyPermissions()` already short-circuits on `"none"` (line 299: `if (mode === "none") return`), so no changes to that function are needed.
|
||||
- **Alternatives rejected:** Fix `from-commands` flattening — impossible to do correctly with global-only target. Remove flag entirely — too restrictive for power users.
|
||||
- **Assumptions:** The `applyPermissions()` function with mode `"none"` leaves `config.permission` and `config.tools` as `undefined`.
|
||||
- **Reversal trigger:** If OpenCode adds per-command permission scoping, `from-commands` could become meaningful again.
|
||||
|
||||
---
|
||||
|
||||
## ADRs To Create
|
||||
|
||||
Create `docs/decisions/` directory (does not exist on `main`). ADRs follow `AGENTS.md` numbering convention: `0001-short-title.md`.
|
||||
|
||||
### ADR 0001: OpenCode commands written as `.md` files, not in `opencode.json`
|
||||
|
||||
- **Context:** OpenCode supports two equivalent formats for custom commands. Writing to `opencode.json` requires overwriting or merging the user's config file. Writing `.md` files is additive and non-destructive.
|
||||
- **Decision:** The OpenCode target always emits commands as individual `.md` files in the `commands/` subdirectory. The `command` key is never written to `opencode.json` by this tool.
|
||||
- **Consequences:**
|
||||
- Positive: Installs are non-destructive. Commands are visible as individual files, easy to inspect. Consistent with agents/skills handling.
|
||||
- Negative: Users inspecting `opencode.json` won't see plugin commands; they must look in `commands/`.
|
||||
- Neutral: Requires OpenCode >= the version with command file support (confirmed stable).
|
||||
|
||||
### ADR 0002: Plugin merges into existing `opencode.json` rather than replacing it
|
||||
|
||||
- **Context:** Users have existing `opencode.json` files with personal configuration. The install command previously backed up and replaced this file entirely, destroying user settings.
|
||||
- **Decision:** `writeOpenCodeBundle` reads existing `opencode.json` (if present), deep-merges plugin-provided keys without overwriting user-set values, and writes the merged result. User keys always win on conflict.
|
||||
- **Consequences:**
|
||||
- Positive: User config preserved across installs. Re-installs are idempotent for user-set values.
|
||||
- Negative: Plugin cannot remove or update an MCP server entry if the user already has one with the same name.
|
||||
- Neutral: Backup of pre-merge file is still created for safety.
|
||||
|
||||
### ADR 0003: Global permissions not written to `opencode.json` by default
|
||||
|
||||
- **Context:** Claude commands carry `allowedTools` as per-command restrictions. OpenCode has no per-command permission mechanism. Writing per-command restrictions as global permissions is semantically incorrect and pollutes the user's global config.
|
||||
- **Decision:** `--permissions` defaults to `"none"`. The plugin never writes `permission` or `tools` to `opencode.json` unless the user explicitly passes `--permissions broad` or `--permissions from-commands`.
|
||||
- **Consequences:**
|
||||
- Positive: User's global OpenCode permissions are never silently modified.
|
||||
- Negative: Users who relied on auto-set permissions must now pass the flag explicitly.
|
||||
- Neutral: The `"broad"` and `"from-commands"` modes still work as documented for opt-in use.
|
||||
|
||||
---
|
||||
|
||||
## Assumptions & Invalidation Triggers
|
||||
|
||||
- **Assumption:** OpenCode command `.md` frontmatter supports `description`, `agent`, `model` and does NOT support `permission` or `tools`.
|
||||
- **If this changes:** The converter could emit per-command permissions in command frontmatter, making `from-commands` mode semantically correct. Phase 2 would need a new code path.
|
||||
|
||||
- **Assumption:** `readJson()` and `pathExists()` exist in `src/utils/files.ts` and work as expected.
|
||||
- **If this changes:** Phase 4's merge logic needs alternative I/O utilities.
|
||||
|
||||
- **Assumption:** `applyPermissions()` with mode `"none"` returns early at line 299 and does not set `config.permission` or `config.tools`.
|
||||
- **If this changes:** The merge logic in Phase 4 might still merge stale data. Verify before implementing.
|
||||
|
||||
- **Assumption:** 180 tests pass on `main` at commit `174cd4c` with `bun test`.
|
||||
- **If this changes:** Do not proceed until the discrepancy is understood.
|
||||
|
||||
- **Assumption:** `formatFrontmatter()` in `src/utils/frontmatter.ts` handles `Record<string, unknown>` data and string body, producing valid YAML frontmatter. It filters out `undefined` values (line 35). It already supports nested objects/arrays via `formatYamlLine()`.
|
||||
- **If this changes:** Phase 2's command file content generation would produce malformed output.
|
||||
|
||||
- **Assumption:** The `backupFile()` function in `src/utils/files.ts` returns `null` if the file does not exist, and returns the backup path if it does. It does NOT throw on missing files.
|
||||
- **If this changes:** Phase 4's backup-before-write for command files would need error handling.
|
||||
|
||||
---
|
||||
|
||||
## Phases
|
||||
|
||||
### Phase 1: Add `OpenCodeCommandFile` type and update `OpenCodeBundle`
|
||||
|
||||
**What:** In `src/types/opencode.ts`:
|
||||
- Add a new type `OpenCodeCommandFile` with `name: string` (command name, used as filename stem) and `content: string` (full file content: YAML frontmatter + body).
|
||||
- Add `commandFiles: OpenCodeCommandFile[]` field to `OpenCodeBundle`.
|
||||
- Remove `command?: Record<string, OpenCodeCommandConfig>` from `OpenCodeConfig`.
|
||||
- Remove the `OpenCodeCommandConfig` type entirely (lines 23-28).
|
||||
|
||||
**Why:** This is the foundational type change that all subsequent phases depend on. Commands move from the config object to individual file entries in the bundle.
|
||||
|
||||
**Test first:**
|
||||
|
||||
File: `tests/converter.test.ts`
|
||||
|
||||
Before making any type changes, update the test file to reflect the new shape. The existing tests will fail because they reference `bundle.config.command` and `OpenCodeBundle` doesn't have `commandFiles` yet.
|
||||
|
||||
Tests to modify (they will fail after type changes, then pass after Phase 2):
|
||||
- `"maps commands, permissions, and agents"` (line 11): Change `bundle.config.command?.["workflows:review"]` to `bundle.commandFiles.find(f => f.name === "workflows:review")`. Change `bundle.config.command?.["plan_review"]` to `bundle.commandFiles.find(f => f.name === "plan_review")`.
|
||||
- `"normalizes models and infers temperature"` (line 60): Change `bundle.config.command?.["workflows:work"]` to check `bundle.commandFiles.find(f => f.name === "workflows:work")` and parse its frontmatter for model.
|
||||
- `"excludes commands with disable-model-invocation from command map"` (line 202): Change `bundle.config.command?.["deploy-docs"]` to `bundle.commandFiles.find(f => f.name === "deploy-docs")`.
|
||||
- `"rewrites .claude/ paths to .opencode/ in command bodies"` (line 217): Change `bundle.config.command?.["review"]?.template` to access `bundle.commandFiles.find(f => f.name === "review")?.content`.
|
||||
|
||||
Also update `tests/opencode-writer.test.ts`:
|
||||
- Add `commandFiles: []` to every `OpenCodeBundle` literal in all 4 existing tests (lines 20, 43, 67, 98). These bundles currently only have `config`, `agents`, `plugins`, `skillDirs`.
|
||||
|
||||
**Implementation:**
|
||||
|
||||
In `src/types/opencode.ts`:
|
||||
1. Remove lines 23-28 (`OpenCodeCommandConfig` type).
|
||||
2. Remove line 10 (`command?: Record<string, OpenCodeCommandConfig>`) from `OpenCodeConfig`.
|
||||
3. Add after line 47:
|
||||
```typescript
|
||||
export type OpenCodeCommandFile = {
|
||||
name: string // command name, used as the filename stem: <name>.md
|
||||
content: string // full file content: YAML frontmatter + body
|
||||
}
|
||||
```
|
||||
4. Add `commandFiles: OpenCodeCommandFile[]` to `OpenCodeBundle` (between `agents` and `plugins`).
|
||||
|
||||
In `src/converters/claude-to-opencode.ts`:
|
||||
- Update the import on line 11: Remove `OpenCodeCommandConfig` from the import. Add `OpenCodeCommandFile`.
|
||||
|
||||
**Code comments required:**
|
||||
- Above the `commandFiles` field in `OpenCodeBundle`: `// Commands are written as individual .md files, not in opencode.json. See ADR-001.`
|
||||
|
||||
**Verification:** `bun test` will show failures in converter tests (they reference the old command format). This is expected — Phase 2 fixes them.
|
||||
|
||||
---
|
||||
|
||||
### Phase 2: Convert `convertCommands()` to emit `.md` command files
|
||||
|
||||
**What:** In `src/converters/claude-to-opencode.ts`:
|
||||
- Rewrite `convertCommands()` (line 114) to return `OpenCodeCommandFile[]` instead of `Record<string, OpenCodeCommandConfig>`.
|
||||
- Each command becomes a `.md` file with YAML frontmatter (`description`, optionally `model`) and body (the template text with Claude path rewriting applied).
|
||||
- In `convertClaudeToOpenCode()` (line 64): replace `commandMap` with `commandFiles`. Remove `config.command` assignment. Add `commandFiles` to returned bundle.
|
||||
|
||||
**Why:** This is the core conversion logic change that implements ADR-001.
|
||||
|
||||
**Test first:**
|
||||
|
||||
File: `tests/converter.test.ts`
|
||||
|
||||
The tests were already updated in Phase 1 to reference `bundle.commandFiles`. Now they need to pass. Specific assertions:
|
||||
|
||||
1. Rename `"maps commands, permissions, and agents"` to `"from-commands mode: maps allowedTools to global permission block"` — to clarify this tests an opt-in mode, not the default.
|
||||
- Assert `bundle.config.command` is `undefined` (it no longer exists on the type, but accessing it returns `undefined`).
|
||||
- Assert `bundle.commandFiles.find(f => f.name === "workflows:review")` is defined.
|
||||
- Assert `bundle.commandFiles.find(f => f.name === "plan_review")` is defined.
|
||||
- Permission assertions remain unchanged (they test `from-commands` mode explicitly).
|
||||
|
||||
2. `"normalizes models and infers temperature"`:
|
||||
- Find `workflows:work` in `bundle.commandFiles`, parse its frontmatter with `parseFrontmatter()`, assert `data.model === "openai/gpt-4o"`.
|
||||
|
||||
3. `"excludes commands with disable-model-invocation from command map"` — rename to `"excludes commands with disable-model-invocation from commandFiles"`:
|
||||
- Assert `bundle.commandFiles.find(f => f.name === "deploy-docs")` is `undefined`.
|
||||
- Assert `bundle.commandFiles.find(f => f.name === "workflows:review")` is defined.
|
||||
|
||||
4. `"rewrites .claude/ paths to .opencode/ in command bodies"`:
|
||||
- Find `review` in `bundle.commandFiles`, assert `content` contains `"compound-engineering.local.md"`.
|
||||
|
||||
5. Add NEW test: `"command .md files include description in frontmatter"`:
|
||||
- Create a minimal `ClaudePlugin` with one command (`name: "test-cmd"`, `description: "Test description"`, `body: "Do the thing"`).
|
||||
- Convert with `permissions: "none"`.
|
||||
- Find the command file, parse frontmatter, assert `data.description === "Test description"`.
|
||||
- Assert the body (after frontmatter) contains `"Do the thing"`.
|
||||
|
||||
**Implementation:**
|
||||
|
||||
In `src/converters/claude-to-opencode.ts`:
|
||||
|
||||
Replace lines 114-128 (`convertCommands` function):
|
||||
```typescript
|
||||
// Commands are written as individual .md files rather than entries in opencode.json.
|
||||
// Chosen over JSON map because opencode resolves commands by filename at runtime (ADR-001).
|
||||
function convertCommands(commands: ClaudeCommand[]): OpenCodeCommandFile[] {
|
||||
const files: OpenCodeCommandFile[] = []
|
||||
for (const command of commands) {
|
||||
if (command.disableModelInvocation) continue
|
||||
const frontmatter: Record<string, unknown> = {
|
||||
description: command.description,
|
||||
}
|
||||
if (command.model && command.model !== "inherit") {
|
||||
frontmatter.model = normalizeModel(command.model)
|
||||
}
|
||||
const content = formatFrontmatter(frontmatter, rewriteClaudePaths(command.body))
|
||||
files.push({ name: command.name, content })
|
||||
}
|
||||
return files
|
||||
}
|
||||
```
|
||||
|
||||
Replace lines 64-87 (`convertClaudeToOpenCode` function body):
|
||||
- Change line 69: `const commandFiles = convertCommands(plugin.commands)`
|
||||
- Change lines 73-77 (config construction): Remove the `command: ...` line. Config should only have `$schema` and `mcp`.
|
||||
- Change line 81-86 (return): Replace `plugins` in the return with `commandFiles, plugins` (add `commandFiles` field to returned bundle).
|
||||
|
||||
**Code comments required:**
|
||||
- Above `convertCommands()`: `// Commands are written as individual .md files rather than entries in opencode.json.` and `// Chosen over JSON map because opencode resolves commands by filename at runtime (ADR-001).`
|
||||
|
||||
**Verification:** Run `bun test tests/converter.test.ts`. All converter tests must pass. Then run `bun test` — writer tests should still fail (they expect the old bundle shape; fixed in Phase 1's test updates) but converter tests pass.
|
||||
|
||||
---
|
||||
|
||||
### Phase 3: Add `commandsDir` to path resolver and write command files
|
||||
|
||||
**What:** In `src/targets/opencode.ts`:
|
||||
- Add `commandsDir` to the return value of `resolveOpenCodePaths()` for both branches (global and custom output dir).
|
||||
- In `writeOpenCodeBundle()`, iterate `bundle.commandFiles` and write each as `<commandsDir>/<name>.md` with backup-before-overwrite.
|
||||
|
||||
**Why:** This creates the file output mechanism for command `.md` files. Separated from Phase 4 (merge logic) for testability.
|
||||
|
||||
**Test first:**
|
||||
|
||||
File: `tests/opencode-writer.test.ts`
|
||||
|
||||
Add these new tests:
|
||||
|
||||
1. `"writes command files as .md in commands/ directory"`:
|
||||
- Create a bundle with one `commandFiles` entry: `{ name: "my-cmd", content: "---\ndescription: Test\n---\n\nDo something." }`.
|
||||
- Use an output root of `path.join(tempRoot, ".config", "opencode")` (global-style).
|
||||
- Assert `exists(path.join(outputRoot, "commands", "my-cmd.md"))` is true.
|
||||
- Read the file, assert content matches (with trailing newline: `content + "\n"`).
|
||||
|
||||
2. `"backs up existing command .md file before overwriting"`:
|
||||
- Pre-create `commands/my-cmd.md` with old content.
|
||||
- Write a bundle with a `commandFiles` entry for `my-cmd`.
|
||||
- Assert a `.bak.` file exists in `commands/` directory.
|
||||
- Assert new content is written.
|
||||
|
||||
**Implementation:**
|
||||
|
||||
In `resolveOpenCodePaths()`:
|
||||
- In the global branch (line 39-46): Add `commandsDir: path.join(outputRoot, "commands")` with comment: `// .md command files; alternative to the command key in opencode.json`
|
||||
- In the custom branch (line 49-56): Add `commandsDir: path.join(outputRoot, ".opencode", "commands")` with same comment.
|
||||
|
||||
In `writeOpenCodeBundle()`:
|
||||
- After the agents loop (line 18), add:
|
||||
```typescript
|
||||
const commandsDir = paths.commandsDir
|
||||
for (const commandFile of bundle.commandFiles) {
|
||||
const dest = path.join(commandsDir, `${commandFile.name}.md`)
|
||||
const cmdBackupPath = await backupFile(dest)
|
||||
if (cmdBackupPath) {
|
||||
console.log(`Backed up existing command file to ${cmdBackupPath}`)
|
||||
}
|
||||
await writeText(dest, commandFile.content + "\n")
|
||||
}
|
||||
```
|
||||
|
||||
**Code comments required:**
|
||||
- Inline comment on `commandsDir` in both `resolveOpenCodePaths` branches: `// .md command files; alternative to the command key in opencode.json`
|
||||
|
||||
**Verification:** Run `bun test tests/opencode-writer.test.ts`. The two new command file tests must pass. Existing tests must still pass (they have `commandFiles: []` from Phase 1 updates).
|
||||
|
||||
---
|
||||
|
||||
### Phase 4: Replace config overwrite with deep-merge
|
||||
|
||||
**What:** In `src/targets/opencode.ts`:
|
||||
- Replace `writeJson(paths.configPath, bundle.config)` (line 13) with a call to a new `mergeOpenCodeConfig()` function.
|
||||
- `mergeOpenCodeConfig()` reads the existing `opencode.json` (if present), merges plugin-provided keys using user-wins-on-conflict strategy, and returns the merged config.
|
||||
- Import `pathExists` and `readJson` from `../utils/files` (add to existing import on line 2).
|
||||
|
||||
**Why:** This implements ADR-002 — the user's existing config is preserved across installs.
|
||||
|
||||
**Test first:**
|
||||
|
||||
File: `tests/opencode-writer.test.ts`
|
||||
|
||||
Modify existing test and add new tests:
|
||||
|
||||
1. Rename `"backs up existing opencode.json before overwriting"` (line 88) to `"merges plugin config into existing opencode.json without destroying user keys"`:
|
||||
- Pre-create `opencode.json` with `{ $schema: "https://opencode.ai/config.json", custom: "value" }`.
|
||||
- Write a bundle with `config: { $schema: "...", mcp: { "plugin-server": { type: "local", command: "uvx", args: ["plugin-srv"] } } }`.
|
||||
- Assert merged config has BOTH `custom: "value"` (user key) AND `mcp["plugin-server"]` (plugin key).
|
||||
- Assert backup file exists with original content.
|
||||
|
||||
2. NEW: `"merges mcp servers without overwriting user entries"`:
|
||||
- Pre-create `opencode.json` with `{ mcp: { "user-server": { type: "local", command: "uvx", args: ["user-srv"] } } }`.
|
||||
- Write a bundle with `config.mcp` containing both `"plugin-server"` (new) and `"user-server"` (conflict — different args).
|
||||
- Assert both servers exist in merged output.
|
||||
- Assert `user-server` keeps user's original args (user wins on conflict).
|
||||
- Assert `plugin-server` is present with plugin's args.
|
||||
|
||||
3. NEW: `"preserves unrelated user keys when merging opencode.json"`:
|
||||
- Pre-create `opencode.json` with `{ model: "my-model", theme: "dark", mcp: {} }`.
|
||||
- Write a bundle with `config: { $schema: "...", mcp: { "plugin-server": ... }, permission: { "bash": "allow" } }`.
|
||||
- Assert `model` and `theme` are preserved.
|
||||
- Assert plugin additions are present.
|
||||
|
||||
**Implementation:**
|
||||
|
||||
Add to imports in `src/targets/opencode.ts` line 2:
|
||||
```typescript
|
||||
import { backupFile, copyDir, ensureDir, pathExists, readJson, writeJson, writeText } from "../utils/files"
|
||||
import type { OpenCodeBundle, OpenCodeConfig } from "../types/opencode"
|
||||
```
|
||||
|
||||
Add `mergeOpenCodeConfig()` function:
|
||||
```typescript
|
||||
async function mergeOpenCodeConfig(
|
||||
configPath: string,
|
||||
incoming: OpenCodeConfig,
|
||||
): Promise<OpenCodeConfig> {
|
||||
// If no existing config, write plugin config as-is
|
||||
if (!(await pathExists(configPath))) return incoming
|
||||
|
||||
let existing: OpenCodeConfig
|
||||
try {
|
||||
existing = await readJson<OpenCodeConfig>(configPath)
|
||||
} catch {
|
||||
// Safety first per AGENTS.md -- do not destroy user data even if their config is malformed.
|
||||
// Warn and fall back to plugin-only config rather than crashing.
|
||||
console.warn(
|
||||
`Warning: existing ${configPath} is not valid JSON. Writing plugin config without merging.`
|
||||
)
|
||||
return incoming
|
||||
}
|
||||
|
||||
// User config wins on conflict -- see ADR-002
|
||||
// MCP servers: add plugin entries, skip keys already in user config.
|
||||
const mergedMcp = {
|
||||
...(incoming.mcp ?? {}),
|
||||
...(existing.mcp ?? {}), // existing takes precedence (overwrites same-named plugin entries)
|
||||
}
|
||||
|
||||
// Permission: add plugin entries, skip keys already in user config.
|
||||
const mergedPermission = incoming.permission
|
||||
? {
|
||||
...(incoming.permission),
|
||||
...(existing.permission ?? {}), // existing takes precedence
|
||||
}
|
||||
: existing.permission
|
||||
|
||||
// Tools: same pattern
|
||||
const mergedTools = incoming.tools
|
||||
? {
|
||||
...(incoming.tools),
|
||||
...(existing.tools ?? {}),
|
||||
}
|
||||
: existing.tools
|
||||
|
||||
return {
|
||||
...existing, // all user keys preserved
|
||||
$schema: incoming.$schema ?? existing.$schema,
|
||||
mcp: Object.keys(mergedMcp).length > 0 ? mergedMcp : undefined,
|
||||
permission: mergedPermission,
|
||||
tools: mergedTools,
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
In `writeOpenCodeBundle()`, replace line 13 (`await writeJson(paths.configPath, bundle.config)`) with:
|
||||
```typescript
|
||||
const merged = await mergeOpenCodeConfig(paths.configPath, bundle.config)
|
||||
await writeJson(paths.configPath, merged)
|
||||
```
|
||||
|
||||
**Code comments required:**
|
||||
- Above `mergeOpenCodeConfig()`: `// Merges plugin config into existing opencode.json. User keys win on conflict. See ADR-002.`
|
||||
- On the `...(existing.mcp ?? {})` line: `// existing takes precedence (overwrites same-named plugin entries)`
|
||||
- On malformed JSON catch: `// Safety first per AGENTS.md -- do not destroy user data even if their config is malformed.`
|
||||
|
||||
**Verification:** Run `bun test tests/opencode-writer.test.ts`. All tests must pass including the renamed test and the 2 new merge tests.
|
||||
|
||||
---
|
||||
|
||||
### Phase 5: Change `--permissions` default to `"none"`
|
||||
|
||||
**What:** In `src/commands/install.ts`, change line 51 `default: "broad"` to `default: "none"`. Update the description string.
|
||||
|
||||
**Why:** This implements ADR-003 — stops polluting user's global config with permissions by default.
|
||||
|
||||
**Test first:**
|
||||
|
||||
File: `tests/cli.test.ts`
|
||||
|
||||
Add these tests:
|
||||
|
||||
1. `"install --to opencode uses permissions:none by default"`:
|
||||
- Run install with no `--permissions` flag against the fixture plugin.
|
||||
- Read the written `opencode.json`.
|
||||
- Assert it does NOT contain a `permission` key.
|
||||
- Assert it does NOT contain a `tools` key.
|
||||
|
||||
2. `"install --to opencode --permissions broad writes permission block"`:
|
||||
- Run install with `--permissions broad` against the fixture plugin.
|
||||
- Read the written `opencode.json`.
|
||||
- Assert it DOES contain a `permission` key with values.
|
||||
|
||||
**Implementation:**
|
||||
|
||||
In `src/commands/install.ts`:
|
||||
- Line 51: Change `default: "broad"` to `default: "none"`.
|
||||
- Line 52: Change description to `"Permission mapping written to opencode.json: none (default) | broad | from-commands"`.
|
||||
|
||||
**Code comments required:**
|
||||
- On the `default: "none"` line: `// Default is "none" -- writing global permissions to opencode.json pollutes user config. See ADR-003.`
|
||||
|
||||
**Verification:** Run `bun test tests/cli.test.ts`. All CLI tests must pass including the 2 new permission tests. Then run `bun test` — all tests (180 original + new ones) must pass.
|
||||
|
||||
---
|
||||
|
||||
### Phase 6: Update `AGENTS.md` and `README.md`
|
||||
|
||||
**What:** Update documentation to reflect all three changes.
|
||||
|
||||
**Why:** Keeps docs accurate for future contributors and users.
|
||||
|
||||
**Test first:** No tests required for documentation changes.
|
||||
|
||||
**Implementation:**
|
||||
|
||||
In `AGENTS.md` line 10, replace:
|
||||
```
|
||||
- **Output Paths:** Keep OpenCode output at `opencode.json` and `.opencode/{agents,skills,plugins}`.
|
||||
```
|
||||
with:
|
||||
```
|
||||
- **Output Paths:** Keep OpenCode output at `opencode.json` and `.opencode/{agents,skills,plugins}`. For OpenCode, commands go to `~/.config/opencode/commands/<name>.md`; `opencode.json` is deep-merged (never overwritten wholesale).
|
||||
```
|
||||
|
||||
In `README.md` line 54, replace:
|
||||
```
|
||||
OpenCode output is written to `~/.config/opencode` by default, with `opencode.json` at the root and `agents/`, `skills/`, and `plugins/` alongside it.
|
||||
```
|
||||
with:
|
||||
```
|
||||
OpenCode output is written to `~/.config/opencode` by default. Commands are written as individual `.md` files to `~/.config/opencode/commands/<name>.md`. Agents, skills, and plugins are written to the corresponding subdirectories alongside. `opencode.json` (MCP servers) is deep-merged into any existing file -- user keys such as `model`, `theme`, and `provider` are preserved, and user values win on conflicts. Command files are backed up before being overwritten.
|
||||
```
|
||||
|
||||
Also update `AGENTS.md` to add a Repository Docs Conventions section if not present:
|
||||
```
|
||||
## Repository Docs Conventions
|
||||
|
||||
- **ADRs** live in `docs/decisions/` and are numbered with 4-digit zero-padding: `0001-short-title.md`, `0002-short-title.md`, etc.
|
||||
- **Orchestrator run reports** live in `docs/reports/`.
|
||||
|
||||
When recording a significant decision (new provider, output format change, merge strategy), create an ADR in `docs/decisions/` following the numbering sequence.
|
||||
```
|
||||
|
||||
**Code comments required:** None.
|
||||
|
||||
**Verification:** Read the updated files and confirm accuracy. Run `bun test` to confirm no regressions.
|
||||
|
||||
---
|
||||
|
||||
## TDD Enforcement
|
||||
|
||||
The executing agent MUST follow this sequence for every phase that touches source code:
|
||||
|
||||
1. Write the test(s) first in the test file.
|
||||
2. Run `bun test <test-file>` and confirm the new/modified tests FAIL (red).
|
||||
3. Implement the code change.
|
||||
4. Run `bun test <test-file>` and confirm the new/modified tests PASS (green).
|
||||
5. Run `bun test` (all tests) and confirm no regressions.
|
||||
|
||||
**Exception:** Phase 6 is documentation only. Run `bun test` after to confirm no regressions but no red/green cycle needed.
|
||||
|
||||
**Note on Phase 1:** Type changes alone will cause test failures. Phase 1 and Phase 2 are tightly coupled — the tests updated in Phase 1 will not pass until Phase 2's implementation is complete. The executing agent should:
|
||||
1. Update tests in Phase 1 (expect them to fail — both due to type errors and logic changes).
|
||||
2. Implement type changes in Phase 1.
|
||||
3. Implement converter changes in Phase 2.
|
||||
4. Confirm all converter tests pass after Phase 2.
|
||||
|
||||
---
|
||||
|
||||
## Constraints
|
||||
|
||||
**Do not modify:**
|
||||
- `src/converters/claude-to-opencode.ts` lines 294-417 (`applyPermissions()`, `normalizeTool()`, `parseToolSpec()`, `normalizePattern()`) — these functions are correct for `"broad"` and `"from-commands"` modes. Only the default that triggers them is changing.
|
||||
- Any files under `tests/fixtures/` — these are data files, not test logic.
|
||||
- `src/types/claude.ts` — no changes to source types.
|
||||
- `src/parsers/claude.ts` — no changes to parser logic.
|
||||
- `src/utils/files.ts` — all needed utilities already exist. Do not add new utility functions.
|
||||
- `src/utils/frontmatter.ts` — already handles the needed formatting.
|
||||
|
||||
**Dependencies not to add:** None. No new npm/bun packages.
|
||||
|
||||
**Patterns to follow:**
|
||||
- Existing writer tests in `tests/opencode-writer.test.ts` use `fs.mkdtemp()` for temp directories and the local `exists()` helper function.
|
||||
- Existing CLI tests in `tests/cli.test.ts` use `Bun.spawn()` to invoke the CLI.
|
||||
- Existing converter tests in `tests/converter.test.ts` use `loadClaudePlugin(fixtureRoot)` for real fixtures and inline `ClaudePlugin` objects for isolated tests.
|
||||
- ADR format: Follow `AGENTS.md` numbering convention `0001-short-title.md` with sections: Status, Date, Context, Decision, Consequences, Plan Reference.
|
||||
- Commits: Use conventional commit format. Reference ADRs in commit bodies.
|
||||
- Branch: Create `feature/opencode-commands-md-merge-permissions` from `main`.
|
||||
|
||||
## Final Checklist
|
||||
|
||||
After all phases complete:
|
||||
- [ ] `bun test` passes all tests (180 original + new ones, 0 fail)
|
||||
- [ ] `docs/decisions/0001-opencode-command-output-format.md` exists
|
||||
- [ ] `docs/decisions/0002-opencode-json-merge-strategy.md` exists
|
||||
- [ ] `docs/decisions/0003-opencode-permissions-default-none.md` exists
|
||||
- [ ] `opencode.json` is never fully overwritten — merge logic confirmed by test
|
||||
- [ ] Commands are written as `.md` files — confirmed by test
|
||||
- [ ] `--permissions` defaults to `"none"` — confirmed by CLI test
|
||||
- [ ] `AGENTS.md` and `README.md` updated to reflect new behavior
|
||||
692
docs/solutions/adding-converter-target-providers.md
Normal file
692
docs/solutions/adding-converter-target-providers.md
Normal file
@@ -0,0 +1,692 @@
|
||||
---
|
||||
title: Adding New Converter Target Providers
|
||||
category: architecture
|
||||
tags: [converter, target-provider, plugin-conversion, multi-platform, pattern]
|
||||
created: 2026-02-23
|
||||
severity: medium
|
||||
component: converter-cli
|
||||
problem_type: best_practice
|
||||
root_cause: architectural_pattern
|
||||
---
|
||||
|
||||
# Adding New Converter Target Providers
|
||||
|
||||
## Problem
|
||||
|
||||
When adding support for a new AI platform (e.g., Devin, Cursor, Copilot), the converter CLI architecture requires consistent implementation across types, converters, writers, CLI integration, and tests. Without documented patterns and learnings, new targets take longer to implement and risk architectural inconsistency.
|
||||
|
||||
## Solution
|
||||
|
||||
The compound-engineering-plugin uses a proven **6-phase target provider pattern** that has been successfully applied to 8 targets:
|
||||
|
||||
1. **OpenCode** (primary target, reference implementation)
|
||||
2. **Codex** (second target, established pattern)
|
||||
3. **Droid/Factory** (workflow/agent conversion)
|
||||
4. **Pi** (MCPorter ecosystem)
|
||||
5. **Gemini CLI** (content transformation patterns)
|
||||
6. **Cursor** (command flattening, rule formats)
|
||||
7. **Copilot** (GitHub native, MCP prefixing)
|
||||
8. **Kiro** (limited MCP support)
|
||||
9. **Devin** (playbook conversion, knowledge entries)
|
||||
|
||||
Each implementation follows this architecture precisely, ensuring consistency and maintainability.
|
||||
|
||||
## Architecture: The 6-Phase Pattern
|
||||
|
||||
### Phase 1: Type Definitions (`src/types/{target}.ts`)
|
||||
|
||||
**Purpose:** Define TypeScript types for the intermediate bundle format
|
||||
|
||||
**Key Pattern:**
|
||||
|
||||
```typescript
|
||||
// Exported bundle type used by converter and writer
|
||||
export type {TargetName}Bundle = {
|
||||
// Component arrays matching the target format
|
||||
agents?: {TargetName}Agent[]
|
||||
commands?: {TargetName}Command[]
|
||||
skillDirs?: {TargetName}SkillDir[]
|
||||
mcpServers?: Record<string, {TargetName}McpServer>
|
||||
// Target-specific fields
|
||||
setup?: string // Instructions file content
|
||||
}
|
||||
|
||||
// Individual component types
|
||||
export type {TargetName}Agent = {
|
||||
name: string
|
||||
content: string // Full file content (with frontmatter if applicable)
|
||||
category?: string // e.g., "agent", "rule", "playbook"
|
||||
meta?: Record<string, unknown> // Target-specific metadata
|
||||
}
|
||||
```
|
||||
|
||||
**Key Learnings:**
|
||||
|
||||
- Always include a `content` field (full file text) rather than decomposed fields — it's simpler and matches how files are written
|
||||
- Use intermediate types for complex sections (e.g., `DevinPlaybookSections` in Devin converter) to make section building independently testable
|
||||
- Avoid target-specific fields in the base bundle unless essential — aim for shared structure across targets
|
||||
- Include a `category` field if the target has file-type variants (agents vs. commands vs. rules)
|
||||
|
||||
**Reference Implementations:**
|
||||
- OpenCode: `src/types/opencode.ts` (command + agent split)
|
||||
- Devin: `src/types/devin.ts` (playbooks + knowledge entries)
|
||||
- Copilot: `src/types/copilot.ts` (agents + skills + MCP)
|
||||
|
||||
---
|
||||
|
||||
### Phase 2: Converter (`src/converters/claude-to-{target}.ts`)
|
||||
|
||||
**Purpose:** Transform Claude Code plugin format → target-specific bundle format
|
||||
|
||||
**Key Pattern:**
|
||||
|
||||
```typescript
|
||||
export type ClaudeTo{Target}Options = ClaudeToOpenCodeOptions // Reuse common options
|
||||
|
||||
export function convertClaudeTo{Target}(
|
||||
plugin: ClaudePlugin,
|
||||
_options: ClaudeTo{Target}Options,
|
||||
): {Target}Bundle {
|
||||
// Pre-scan: build maps for cross-reference resolution (agents, commands)
|
||||
// Needed if target requires deduplication or reference tracking
|
||||
const refMap: Record<string, string> = {}
|
||||
for (const agent of plugin.agents) {
|
||||
refMap[normalize(agent.name)] = macroName(agent.name)
|
||||
}
|
||||
|
||||
// Phase 1: Convert agents
|
||||
const agents = plugin.agents.map(a => convert{Target}Agent(a, usedNames, refMap))
|
||||
|
||||
// Phase 2: Convert commands (may depend on agent names for dedup)
|
||||
const commands = plugin.commands.map(c => convert{Target}Command(c, usedNames, refMap))
|
||||
|
||||
// Phase 3: Handle skills (usually pass-through, sometimes conversion)
|
||||
const skillDirs = plugin.skills.map(s => ({ name: s.name, sourceDir: s.sourceDir }))
|
||||
|
||||
// Phase 4: Convert MCP servers (target-specific prefixing/type mapping)
|
||||
const mcpConfig = convertMcpServers(plugin.mcpServers)
|
||||
|
||||
// Phase 5: Warn on unsupported features
|
||||
if (plugin.hooks && Object.keys(plugin.hooks.hooks).length > 0) {
|
||||
console.warn("Warning: {Target} does not support hooks. Hooks were skipped.")
|
||||
}
|
||||
|
||||
return { agents, commands, skillDirs, mcpConfig }
|
||||
}
|
||||
```
|
||||
|
||||
**Content Transformation (`transformContentFor{Target}`):**
|
||||
|
||||
Applied to both agent bodies and command bodies to rewrite paths, command references, and agent mentions:
|
||||
|
||||
```typescript
|
||||
export function transformContentFor{Target}(body: string): string {
|
||||
let result = body
|
||||
|
||||
// 1. Rewrite paths (.claude/ → .github/, ~/.claude/ → ~/.{target}/)
|
||||
result = result
|
||||
.replace(/~\/\.claude\//g, `~/.${targetDir}/`)
|
||||
.replace(/\.claude\//g, `.${targetDir}/`)
|
||||
|
||||
// 2. Transform Task agent calls (to natural language)
|
||||
const taskPattern = /Task\s+([a-z][a-z0-9-]*)\(([^)]+)\)/gm
|
||||
result = result.replace(taskPattern, (_match, agentName: string, args: string) => {
|
||||
const skillName = normalize(agentName)
|
||||
return `Use the ${skillName} skill to: ${args.trim()}`
|
||||
})
|
||||
|
||||
// 3. Flatten slash commands (/workflows:plan → /plan)
|
||||
const slashPattern = /(?<![:\w])\/([a-z][a-z0-9_:-]*?)(?=[\s,."')\]}`]|$)/gi
|
||||
result = result.replace(slashPattern, (match, commandName: string) => {
|
||||
if (commandName.includes("/")) return match // Skip file paths
|
||||
const normalized = normalize(commandName)
|
||||
return `/${normalized}`
|
||||
})
|
||||
|
||||
// 4. Transform @agent-name references
|
||||
const agentPattern = /@([a-z][a-z0-9-]*-(?:agent|reviewer|analyst|...))/gi
|
||||
result = result.replace(agentPattern, (_match, agentName: string) => {
|
||||
return `the ${normalize(agentName)} agent` // or "rule", "playbook", etc.
|
||||
})
|
||||
|
||||
// 5. Remove examples (if target doesn't support them)
|
||||
result = result.replace(/<examples>[\s\S]*?<\/examples>/g, "")
|
||||
|
||||
return result
|
||||
}
|
||||
```
|
||||
|
||||
**Deduplication Pattern (`uniqueName`):**
|
||||
|
||||
Used when target has flat namespaces (Cursor, Copilot, Devin) or when name collisions occur:
|
||||
|
||||
```typescript
|
||||
function uniqueName(base: string, used: Set<string>): string {
|
||||
if (!used.has(base)) {
|
||||
used.add(base)
|
||||
return base
|
||||
}
|
||||
let index = 2
|
||||
while (used.has(`${base}-${index}`)) {
|
||||
index += 1
|
||||
}
|
||||
const name = `${base}-${index}`
|
||||
used.add(name)
|
||||
return name
|
||||
}
|
||||
|
||||
function normalizeName(value: string): string {
|
||||
const trimmed = value.trim()
|
||||
if (!trimmed) return "item"
|
||||
const normalized = trimmed
|
||||
.toLowerCase()
|
||||
.replace(/[\\/]+/g, "-")
|
||||
.replace(/[:\s]+/g, "-")
|
||||
.replace(/[^a-z0-9_-]+/g, "-")
|
||||
.replace(/-+/g, "-")
|
||||
.replace(/^-+|-+$/g, "")
|
||||
return normalized || "item"
|
||||
}
|
||||
|
||||
// Flatten: drops namespace prefix (workflows:plan → plan)
|
||||
function flattenCommandName(name: string): string {
|
||||
const normalized = normalizeName(name)
|
||||
return normalized.replace(/^[a-z]+-/, "") // Drop prefix before first dash
|
||||
}
|
||||
```
|
||||
|
||||
**Key Learnings:**
|
||||
|
||||
1. **Pre-scan for cross-references** — If target requires reference names (macros, URIs, IDs), build a map before conversion. Example: Devin needs macro names like `agent_kieran_rails_reviewer`, so pre-scan builds the map.
|
||||
|
||||
2. **Content transformation is fragile** — Test extensively. Patterns that work for slash commands might false-match on file paths. Use negative lookahead to skip `/etc`, `/usr`, `/var`, etc.
|
||||
|
||||
3. **Simplify heuristics, trust structural mapping** — Don't try to parse agent body for "You are..." or "NEVER do..." patterns. Instead, map agent.description → Overview, agent.body → Procedure, agent.capabilities → Specifications. Heuristics fail on edge cases and are hard to test.
|
||||
|
||||
4. **Normalize early and consistently** — Use the same `normalizeName()` function throughout. Inconsistent normalization causes deduplication bugs.
|
||||
|
||||
5. **MCP servers need target-specific handling:**
|
||||
- **OpenCode:** Merge into `opencode.json` (preserve user keys)
|
||||
- **Copilot:** Prefix env vars with `COPILOT_MCP_`, emit JSON
|
||||
- **Devin:** Write setup instructions file (config is via web UI)
|
||||
- **Cursor:** Pass through as-is
|
||||
|
||||
6. **Warn on unsupported features** — Hooks, Gemini extensions, Kiro-incompatible MCP types. Emit to stderr and continue conversion.
|
||||
|
||||
**Reference Implementations:**
|
||||
- OpenCode: `src/converters/claude-to-opencode.ts` (most comprehensive)
|
||||
- Devin: `src/converters/claude-to-devin.ts` (content transformation + cross-references)
|
||||
- Copilot: `src/converters/claude-to-copilot.ts` (MCP prefixing pattern)
|
||||
|
||||
---
|
||||
|
||||
### Phase 3: Writer (`src/targets/{target}.ts`)
|
||||
|
||||
**Purpose:** Write converted bundle to disk in target-specific directory structure
|
||||
|
||||
**Key Pattern:**
|
||||
|
||||
```typescript
|
||||
export async function write{Target}Bundle(outputRoot: string, bundle: {Target}Bundle): Promise<void> {
|
||||
const paths = resolve{Target}Paths(outputRoot)
|
||||
await ensureDir(paths.root)
|
||||
|
||||
// Write each component type
|
||||
if (bundle.agents?.length > 0) {
|
||||
const agentsDir = path.join(paths.root, "agents")
|
||||
for (const agent of bundle.agents) {
|
||||
await writeText(path.join(agentsDir, `${agent.name}.ext`), agent.content + "\n")
|
||||
}
|
||||
}
|
||||
|
||||
if (bundle.commands?.length > 0) {
|
||||
const commandsDir = path.join(paths.root, "commands")
|
||||
for (const command of bundle.commands) {
|
||||
await writeText(path.join(commandsDir, `${command.name}.ext`), command.content + "\n")
|
||||
}
|
||||
}
|
||||
|
||||
// Copy skills (pass-through case)
|
||||
if (bundle.skillDirs?.length > 0) {
|
||||
const skillsDir = path.join(paths.root, "skills")
|
||||
for (const skill of bundle.skillDirs) {
|
||||
await copyDir(skill.sourceDir, path.join(skillsDir, skill.name))
|
||||
}
|
||||
}
|
||||
|
||||
// Write generated skills (converted from commands)
|
||||
if (bundle.generatedSkills?.length > 0) {
|
||||
const skillsDir = path.join(paths.root, "skills")
|
||||
for (const skill of bundle.generatedSkills) {
|
||||
await writeText(path.join(skillsDir, skill.name, "SKILL.md"), skill.content + "\n")
|
||||
}
|
||||
}
|
||||
|
||||
// Write MCP config (target-specific location and format)
|
||||
if (bundle.mcpServers && Object.keys(bundle.mcpServers).length > 0) {
|
||||
const mcpPath = path.join(paths.root, "mcp.json") // or copilot-mcp-config.json, etc.
|
||||
const backupPath = await backupFile(mcpPath)
|
||||
if (backupPath) {
|
||||
console.log(`Backed up existing MCP config to ${backupPath}`)
|
||||
}
|
||||
await writeJson(mcpPath, { mcpServers: bundle.mcpServers })
|
||||
}
|
||||
|
||||
// Write instructions or setup guides
|
||||
if (bundle.setupInstructions) {
|
||||
const setupPath = path.join(paths.root, "setup-instructions.md")
|
||||
await writeText(setupPath, bundle.setupInstructions + "\n")
|
||||
}
|
||||
}
|
||||
|
||||
// Avoid double-nesting (.target/.target/)
|
||||
function resolve{Target}Paths(outputRoot: string) {
|
||||
const base = path.basename(outputRoot)
|
||||
// If already pointing at .target, write directly into it
|
||||
if (base === ".target") {
|
||||
return { root: outputRoot }
|
||||
}
|
||||
// Otherwise nest under .target
|
||||
return { root: path.join(outputRoot, ".target") }
|
||||
}
|
||||
```
|
||||
|
||||
**Backup Pattern (MCP configs only):**
|
||||
|
||||
MCP configs are often pre-existing and user-edited. Backup before overwrite:
|
||||
|
||||
```typescript
|
||||
// From src/utils/files.ts
|
||||
export async function backupFile(filePath: string): Promise<string | null> {
|
||||
if (!existsSync(filePath)) return null
|
||||
const timestamp = new Date().toISOString().replace(/[:.]/g, "-")
|
||||
const dirname = path.dirname(filePath)
|
||||
const basename = path.basename(filePath)
|
||||
const ext = path.extname(basename)
|
||||
const name = basename.slice(0, -ext.length)
|
||||
const backupPath = path.join(dirname, `${name}.${timestamp}${ext}`)
|
||||
await copyFile(filePath, backupPath)
|
||||
return backupPath
|
||||
}
|
||||
```
|
||||
|
||||
**Key Learnings:**
|
||||
|
||||
1. **Always check for double-nesting** — If output root is already `.target`, don't nest again. Pattern:
|
||||
```typescript
|
||||
if (path.basename(outputRoot) === ".target") {
|
||||
return { root: outputRoot } // Write directly
|
||||
}
|
||||
return { root: path.join(outputRoot, ".target") } // Nest
|
||||
```
|
||||
|
||||
2. **Use `writeText` and `writeJson` helpers** — These handle directory creation and line endings consistently
|
||||
|
||||
3. **Backup MCP configs before overwriting** — MCP JSON files are often hand-edited. Always backup with timestamp.
|
||||
|
||||
4. **Empty bundles should succeed gracefully** — Don't fail if a component array is empty. Many plugins may have no commands or no skills.
|
||||
|
||||
5. **File extensions matter** — Match target conventions exactly:
|
||||
- Copilot: `.agent.md` (note the dot)
|
||||
- Cursor: `.mdc` for rules
|
||||
- Devin: `.devin.md` for playbooks
|
||||
- OpenCode: `.md` for commands
|
||||
|
||||
6. **Permissions for sensitive files** — MCP config with API keys should use `0o600`:
|
||||
```typescript
|
||||
await writeJson(mcpPath, config, { mode: 0o600 })
|
||||
```
|
||||
|
||||
**Reference Implementations:**
|
||||
- Droid: `src/targets/droid.ts` (simpler pattern, good for learning)
|
||||
- Copilot: `src/targets/copilot.ts` (double-nesting pattern)
|
||||
- Devin: `src/targets/devin.ts` (setup instructions file)
|
||||
|
||||
---
|
||||
|
||||
### Phase 4: CLI Wiring
|
||||
|
||||
**File: `src/targets/index.ts`**
|
||||
|
||||
Register the new target in the global target registry:
|
||||
|
||||
```typescript
|
||||
import { convertClaudeTo{Target} } from "../converters/claude-to-{target}"
|
||||
import { write{Target}Bundle } from "./{target}"
|
||||
import type { {Target}Bundle } from "../types/{target}"
|
||||
|
||||
export const targets: Record<string, TargetHandler<any>> = {
|
||||
// ... existing targets ...
|
||||
{target}: {
|
||||
name: "{target}",
|
||||
implemented: true,
|
||||
convert: convertClaudeTo{Target} as TargetHandler<{Target}Bundle>["convert"],
|
||||
write: write{Target}Bundle as TargetHandler<{Target}Bundle>["write"],
|
||||
},
|
||||
}
|
||||
```
|
||||
|
||||
**File: `src/commands/convert.ts` and `src/commands/install.ts`**
|
||||
|
||||
Add output root resolution:
|
||||
|
||||
```typescript
|
||||
// In resolveTargetOutputRoot()
|
||||
if (targetName === "{target}") {
|
||||
return path.join(outputRoot, ".{target}")
|
||||
}
|
||||
|
||||
// Update --to flag description
|
||||
const toDescription = "Target format (opencode | codex | droid | cursor | copilot | kiro | {target})"
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
### Phase 5: Sync Support (Optional)
|
||||
|
||||
**File: `src/sync/{target}.ts`**
|
||||
|
||||
If the target supports syncing personal skills and MCP servers:
|
||||
|
||||
```typescript
|
||||
export async function syncTo{Target}(outputRoot: string): Promise<void> {
|
||||
const personalSkillsDir = path.join(expandHome("~/.claude/skills"))
|
||||
const personalSettings = loadSettings(expandHome("~/.claude/settings.json"))
|
||||
|
||||
const skillsDest = path.join(outputRoot, ".{target}", "skills")
|
||||
await ensureDir(skillsDest)
|
||||
|
||||
// Symlink personal skills
|
||||
if (existsSync(personalSkillsDir)) {
|
||||
const skills = readdirSync(personalSkillsDir)
|
||||
for (const skill of skills) {
|
||||
if (!isValidSkillName(skill)) continue
|
||||
const source = path.join(personalSkillsDir, skill)
|
||||
const dest = path.join(skillsDest, skill)
|
||||
await forceSymlink(source, dest)
|
||||
}
|
||||
}
|
||||
|
||||
// Merge MCP servers if applicable
|
||||
if (personalSettings.mcpServers) {
|
||||
const mcpPath = path.join(outputRoot, ".{target}", "mcp.json")
|
||||
const existing = readJson(mcpPath) || {}
|
||||
const merged = {
|
||||
...existing,
|
||||
mcpServers: {
|
||||
...existing.mcpServers,
|
||||
...personalSettings.mcpServers,
|
||||
},
|
||||
}
|
||||
await writeJson(mcpPath, merged, { mode: 0o600 })
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
**File: `src/commands/sync.ts`**
|
||||
|
||||
```typescript
|
||||
// Add to validTargets array
|
||||
const validTargets = ["opencode", "codex", "droid", "cursor", "pi", "{target}"] as const
|
||||
|
||||
// In resolveOutputRoot()
|
||||
case "{target}":
|
||||
return path.join(process.cwd(), ".{target}")
|
||||
|
||||
// In main switch
|
||||
case "{target}":
|
||||
await syncTo{Target}(outputRoot)
|
||||
break
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
### Phase 6: Tests
|
||||
|
||||
**File: `tests/{target}-converter.test.ts`**
|
||||
|
||||
Test converter using inline `ClaudePlugin` fixtures:
|
||||
|
||||
```typescript
|
||||
describe("convertClaudeTo{Target}", () => {
|
||||
it("converts agents to {target} format", () => {
|
||||
const plugin: ClaudePlugin = {
|
||||
name: "test",
|
||||
agents: [
|
||||
{
|
||||
name: "test-agent",
|
||||
description: "Test description",
|
||||
body: "Test body",
|
||||
capabilities: ["Cap 1", "Cap 2"],
|
||||
},
|
||||
],
|
||||
commands: [],
|
||||
skills: [],
|
||||
}
|
||||
|
||||
const bundle = convertClaudeTo{Target}(plugin, {})
|
||||
|
||||
expect(bundle.agents).toHaveLength(1)
|
||||
expect(bundle.agents[0].name).toBe("test-agent")
|
||||
expect(bundle.agents[0].content).toContain("Test description")
|
||||
})
|
||||
|
||||
it("normalizes agent names", () => {
|
||||
const plugin: ClaudePlugin = {
|
||||
name: "test",
|
||||
agents: [
|
||||
{ name: "Test Agent", description: "", body: "", capabilities: [] },
|
||||
],
|
||||
commands: [],
|
||||
skills: [],
|
||||
}
|
||||
|
||||
const bundle = convertClaudeTo{Target}(plugin, {})
|
||||
expect(bundle.agents[0].name).toBe("test-agent")
|
||||
})
|
||||
|
||||
it("deduplicates colliding names", () => {
|
||||
const plugin: ClaudePlugin = {
|
||||
name: "test",
|
||||
agents: [
|
||||
{ name: "Agent Name", description: "", body: "", capabilities: [] },
|
||||
{ name: "Agent Name", description: "", body: "", capabilities: [] },
|
||||
],
|
||||
commands: [],
|
||||
skills: [],
|
||||
}
|
||||
|
||||
const bundle = convertClaudeTo{Target}(plugin, {})
|
||||
expect(bundle.agents.map(a => a.name)).toEqual(["agent-name", "agent-name-2"])
|
||||
})
|
||||
|
||||
it("transforms content paths (.claude → .{target})", () => {
|
||||
const result = transformContentFor{Target}("See ~/.claude/config")
|
||||
expect(result).toContain("~/.{target}/config")
|
||||
})
|
||||
|
||||
it("warns when hooks are present", () => {
|
||||
const spy = jest.spyOn(console, "warn")
|
||||
const plugin: ClaudePlugin = {
|
||||
name: "test",
|
||||
agents: [],
|
||||
commands: [],
|
||||
skills: [],
|
||||
hooks: { hooks: { "file:save": "test" } },
|
||||
}
|
||||
|
||||
convertClaudeTo{Target}(plugin, {})
|
||||
expect(spy).toHaveBeenCalledWith(expect.stringContaining("hooks"))
|
||||
})
|
||||
})
|
||||
```
|
||||
|
||||
**File: `tests/{target}-writer.test.ts`**
|
||||
|
||||
Test writer using temp directories (from `tmp` package):
|
||||
|
||||
```typescript
|
||||
describe("write{Target}Bundle", () => {
|
||||
it("writes agents to {target} format", async () => {
|
||||
const tmpDir = await tmp.dir()
|
||||
const bundle: {Target}Bundle = {
|
||||
agents: [{ name: "test", content: "# Test\nBody" }],
|
||||
commands: [],
|
||||
skillDirs: [],
|
||||
}
|
||||
|
||||
await write{Target}Bundle(tmpDir.path, bundle)
|
||||
|
||||
const written = readFileSync(path.join(tmpDir.path, ".{target}", "agents", "test.ext"), "utf-8")
|
||||
expect(written).toContain("# Test")
|
||||
})
|
||||
|
||||
it("does not double-nest when output root is .{target}", async () => {
|
||||
const tmpDir = await tmp.dir()
|
||||
const targetDir = path.join(tmpDir.path, ".{target}")
|
||||
await ensureDir(targetDir)
|
||||
|
||||
const bundle: {Target}Bundle = {
|
||||
agents: [{ name: "test", content: "# Test" }],
|
||||
commands: [],
|
||||
skillDirs: [],
|
||||
}
|
||||
|
||||
await write{Target}Bundle(targetDir, bundle)
|
||||
|
||||
// Should write to targetDir directly, not targetDir/.{target}
|
||||
const written = path.join(targetDir, "agents", "test.ext")
|
||||
expect(existsSync(written)).toBe(true)
|
||||
})
|
||||
|
||||
it("backs up existing MCP config", async () => {
|
||||
const tmpDir = await tmp.dir()
|
||||
const mcpPath = path.join(tmpDir.path, ".{target}", "mcp.json")
|
||||
await ensureDir(path.dirname(mcpPath))
|
||||
await writeJson(mcpPath, { existing: true })
|
||||
|
||||
const bundle: {Target}Bundle = {
|
||||
agents: [],
|
||||
commands: [],
|
||||
skillDirs: [],
|
||||
mcpServers: { "test": { command: "test" } },
|
||||
}
|
||||
|
||||
await write{Target}Bundle(tmpDir.path, bundle)
|
||||
|
||||
// Backup should exist
|
||||
const backups = readdirSync(path.dirname(mcpPath)).filter(f => f.includes("mcp") && f.includes("-"))
|
||||
expect(backups.length).toBeGreaterThan(0)
|
||||
})
|
||||
})
|
||||
```
|
||||
|
||||
**Key Testing Patterns:**
|
||||
|
||||
- Test normalization, deduplication, content transformation separately
|
||||
- Use inline plugin fixtures (not file-based)
|
||||
- For writer tests, use temp directories and verify file existence
|
||||
- Test edge cases: empty names, empty bodies, special characters
|
||||
- Test error handling: missing files, permission issues
|
||||
|
||||
---
|
||||
|
||||
## Documentation Requirements
|
||||
|
||||
**File: `docs/specs/{target}.md`**
|
||||
|
||||
Document the target format specification:
|
||||
|
||||
- Last verified date (link to official docs)
|
||||
- Config file locations (project-level vs. user-level)
|
||||
- Agent/command/skill format with field descriptions
|
||||
- MCP configuration structure
|
||||
- Character limits (if any)
|
||||
- Example file
|
||||
|
||||
**File: `README.md`**
|
||||
|
||||
Add to supported targets list and include usage examples.
|
||||
|
||||
---
|
||||
|
||||
## Common Pitfalls and Solutions
|
||||
|
||||
| Pitfall | Solution |
|
||||
|---------|----------|
|
||||
| **Double-nesting** (`.cursor/.cursor/`) | Check `path.basename(outputRoot)` before nesting |
|
||||
| **Inconsistent name normalization** | Use single `normalizeName()` function everywhere |
|
||||
| **Fragile content transformation** | Test regex patterns against edge cases (file paths, URLs) |
|
||||
| **Heuristic section extraction fails** | Use structural mapping (description → Overview, body → Procedure) instead |
|
||||
| **MCP config overwrites user edits** | Always backup with timestamp before overwriting |
|
||||
| **Skill body not loaded** | Verify `ClaudeSkill` has `skillPath` field for file reading |
|
||||
| **Missing deduplication** | Build `usedNames` set before conversion, pass to each converter |
|
||||
| **Unsupported features cause silent loss** | Always warn to stderr (hooks, incompatible MCP types, etc.) |
|
||||
| **Test isolation failures** | Use unique temp directories per test, clean up afterward |
|
||||
| **Command namespace collisions after flattening** | Use `uniqueName()` with deduplication, test multiple collisions |
|
||||
|
||||
---
|
||||
|
||||
## Checklist for Adding a New Target
|
||||
|
||||
Use this checklist when adding a new target provider:
|
||||
|
||||
### Implementation
|
||||
- [ ] Create `src/types/{target}.ts` with bundle and component types
|
||||
- [ ] Implement `src/converters/claude-to-{target}.ts` with converter and content transformer
|
||||
- [ ] Implement `src/targets/{target}.ts` with writer
|
||||
- [ ] Register target in `src/targets/index.ts`
|
||||
- [ ] Update `src/commands/convert.ts` (add output root resolution, update help text)
|
||||
- [ ] Update `src/commands/install.ts` (same as convert.ts)
|
||||
- [ ] (Optional) Implement `src/sync/{target}.ts` and update `src/commands/sync.ts`
|
||||
|
||||
### Testing
|
||||
- [ ] Create `tests/{target}-converter.test.ts` with converter tests
|
||||
- [ ] Create `tests/{target}-writer.test.ts` with writer tests
|
||||
- [ ] (Optional) Create `tests/sync-{target}.test.ts` with sync tests
|
||||
- [ ] Run full test suite: `bun test`
|
||||
- [ ] Manual test: `bun run src/index.ts convert --to {target} ./plugins/compound-engineering`
|
||||
|
||||
### Documentation
|
||||
- [ ] Create `docs/specs/{target}.md` with format specification
|
||||
- [ ] Update `README.md` with target in list and usage examples
|
||||
- [ ] Do not hand-add release notes; release automation owns GitHub release notes and release-owned versions
|
||||
|
||||
### Version Bumping
|
||||
- [ ] Use a conventional `feat:` or `fix:` title so release automation can infer the right bump
|
||||
- [ ] Do not hand-start or hand-bump release-owned version lines in `package.json` or plugin manifests
|
||||
- [ ] Run `bun run release:validate` if component counts or descriptions changed
|
||||
|
||||
---
|
||||
|
||||
## References
|
||||
|
||||
### Implementation Examples
|
||||
|
||||
**Reference implementations by priority (easiest to hardest):**
|
||||
|
||||
1. **Droid** (`src/targets/droid.ts`, `src/converters/claude-to-droid.ts`) — Simplest pattern, good learning baseline
|
||||
2. **Copilot** (`src/targets/copilot.ts`, `src/converters/claude-to-copilot.ts`) — MCP prefixing, double-nesting guard
|
||||
3. **Devin** (`src/converters/claude-to-devin.ts`) — Content transformation, cross-references, intermediate types
|
||||
4. **OpenCode** (`src/converters/claude-to-opencode.ts`) — Most comprehensive, handles command structure and config merging
|
||||
|
||||
### Key Utilities
|
||||
|
||||
- `src/utils/frontmatter.ts` — `formatFrontmatter()` and `parseFrontmatter()`
|
||||
- `src/utils/files.ts` — `writeText()`, `writeJson()`, `copyDir()`, `backupFile()`, `ensureDir()`
|
||||
- `src/utils/resolve-home.ts` — `expandHome()` for `~/.{target}` path resolution
|
||||
|
||||
### Existing Tests
|
||||
|
||||
- `tests/cursor-converter.test.ts` — Comprehensive converter tests
|
||||
- `tests/copilot-writer.test.ts` — Writer tests with temp directories
|
||||
- `tests/sync-copilot.test.ts` — Sync pattern with symlinks and config merge
|
||||
|
||||
---
|
||||
|
||||
## Related Files
|
||||
|
||||
- `plugins/compound-engineering/.claude-plugin/plugin.json` — Version and component counts
|
||||
- `CHANGELOG.md` — Pointer to canonical GitHub release history
|
||||
- `README.md` — Usage examples for all targets
|
||||
- `docs/solutions/plugin-versioning-requirements.md` — Checklist for releases
|
||||
152
docs/solutions/codex-skill-prompt-entrypoints.md
Normal file
152
docs/solutions/codex-skill-prompt-entrypoints.md
Normal file
@@ -0,0 +1,152 @@
|
||||
---
|
||||
title: Codex Conversion Skills, Prompts, and Canonical Entry Points
|
||||
category: architecture
|
||||
tags: [codex, converter, skills, prompts, workflows, deprecation]
|
||||
created: 2026-03-15
|
||||
severity: medium
|
||||
component: codex-target
|
||||
problem_type: best_practice
|
||||
root_cause: outdated_target_model
|
||||
---
|
||||
|
||||
# Codex Conversion Skills, Prompts, and Canonical Entry Points
|
||||
|
||||
## Problem
|
||||
|
||||
The Codex target had two conflicting assumptions:
|
||||
|
||||
1. Compound workflow entrypoints like `ce:brainstorm` and `ce:plan` were treated in docs as slash-command-style surfaces.
|
||||
2. The Codex converter installed those entries as copied skills, not as generated prompts.
|
||||
|
||||
That created an inconsistent runtime for cross-workflow handoffs. Copied skill content still contained Claude-style references like `/ce:plan`, but no Codex-native translation was applied to copied `SKILL.md` files, and there was no clear canonical Codex entrypoint model for those workflow skills.
|
||||
|
||||
## What We Learned
|
||||
|
||||
### 1. Codex supports both skills and prompts, and they are different surfaces
|
||||
|
||||
- Skills are loaded from skill roots such as `~/.codex/skills`, and newer Codex code also supports `.agents/skills`.
|
||||
- Prompts are a separate explicit entrypoint surface under `.codex/prompts`.
|
||||
- A skill is not automatically a prompt, and a prompt is not automatically a skill.
|
||||
|
||||
For this repo, that means a copied skill like `ce:plan` is only a skill unless the converter also generates a prompt wrapper for it.
|
||||
|
||||
### 2. Codex skill names come from the directory name
|
||||
|
||||
Codex derives the skill name from the skill directory basename, not from our normalized hyphenated converter name.
|
||||
|
||||
Implication:
|
||||
|
||||
- `~/.codex/skills/ce:plan` loads as the skill `ce:plan`
|
||||
- Rewriting that to `ce-plan` is wrong for skill-to-skill references
|
||||
|
||||
### 3. The original bug was structural, not just wording
|
||||
|
||||
The issue was not that `ce:brainstorm` needed slightly different prose. The real problem was:
|
||||
|
||||
- copied skills bypassed Codex-specific transformation
|
||||
- workflow handoffs referenced a surface that was not clearly represented in installed Codex artifacts
|
||||
|
||||
### 4. Deprecated `workflows:*` aliases add noise in Codex
|
||||
|
||||
The `workflows:*` names exist only for backward compatibility in Claude.
|
||||
|
||||
Copying them into Codex would:
|
||||
|
||||
- duplicate user-facing entrypoints
|
||||
- complicate handoff rewriting
|
||||
- increase ambiguity around which name is canonical
|
||||
|
||||
For Codex, the simpler model is to treat `ce:*` as the only canonical workflow namespace and omit `workflows:*` aliases from installed output.
|
||||
|
||||
## Recommended Codex Model
|
||||
|
||||
Use a two-layer mapping for workflow entrypoints:
|
||||
|
||||
1. **Skills remain the implementation units**
|
||||
- Copy the canonical workflow skills using their exact names, such as `ce:plan`
|
||||
- Preserve exact skill names for any Codex skill references
|
||||
|
||||
2. **Prompts are the explicit entrypoint layer**
|
||||
- Generate prompt wrappers for canonical user-facing workflow entrypoints
|
||||
- Use Codex-safe prompt slugs such as `ce-plan`, `ce-work`, `ce-review`
|
||||
- Prompt wrappers delegate to the exact underlying skill name, such as `ce:plan`
|
||||
|
||||
This gives Codex one clear manual invocation surface while preserving the real loaded skill names internally.
|
||||
|
||||
## Rewrite Rules
|
||||
|
||||
When converting copied `SKILL.md` content for Codex:
|
||||
|
||||
- References to canonical workflow entrypoints should point to generated prompt wrappers
|
||||
- `/ce:plan` -> `/prompts:ce-plan`
|
||||
- `/ce:work` -> `/prompts:ce-work`
|
||||
- References to deprecated aliases should canonicalize to the modern `ce:*` prompt
|
||||
- `/workflows:plan` -> `/prompts:ce-plan`
|
||||
- References to non-entrypoint skills should use the exact skill name, not a normalized alias
|
||||
- Actual Claude commands that are converted to Codex prompts can continue using `/prompts:...`
|
||||
|
||||
### Regression hardening
|
||||
|
||||
When rewriting copied `SKILL.md` files, only known workflow and command references should be rewritten.
|
||||
|
||||
Do not rewrite arbitrary slash-shaped text such as:
|
||||
|
||||
- application routes like `/users` or `/settings`
|
||||
- API path segments like `/state` or `/ops`
|
||||
- URLs such as `https://www.proofeditor.ai/...`
|
||||
|
||||
Unknown slash references should remain unchanged in copied skill content. Otherwise Codex installs silently corrupt unrelated skills while trying to canonicalize workflow handoffs.
|
||||
|
||||
Personal skills loaded from `~/.claude/skills` also need tolerant metadata parsing:
|
||||
|
||||
- malformed YAML frontmatter should not cause the entire skill to disappear
|
||||
- keep the directory name as the stable skill name
|
||||
- treat frontmatter metadata as best-effort only
|
||||
|
||||
## Future Entry Points
|
||||
|
||||
Do not hard-code an allowlist of workflow names in the converter.
|
||||
|
||||
Instead, use a stable rule:
|
||||
|
||||
- `ce:*` = canonical workflow entrypoint
|
||||
- auto-generate a prompt wrapper
|
||||
- `workflows:*` = deprecated alias
|
||||
- omit from Codex output
|
||||
- rewrite references to the canonical `ce:*` target
|
||||
- non-`ce:*` skills = skill-only by default
|
||||
- if a non-`ce:*` skill should also be a prompt entrypoint, mark it explicitly with Codex-specific metadata
|
||||
|
||||
This means future skills like `ce:ideate` should work without manual converter changes.
|
||||
|
||||
## Implementation Guidance
|
||||
|
||||
For the Codex target:
|
||||
|
||||
1. Parse enough skill frontmatter to distinguish command-like entrypoint skills from background skills
|
||||
2. Filter deprecated `workflows:*` alias skills out of Codex installation
|
||||
3. Generate prompt wrappers for canonical `ce:*` workflow skills
|
||||
4. Apply Codex-specific transformation to copied `SKILL.md` files
|
||||
5. Preserve exact Codex skill names internally
|
||||
6. Update README language so Codex entrypoints are documented as Codex-native surfaces, not assumed to be identical to Claude slash commands
|
||||
|
||||
## Prevention
|
||||
|
||||
Before changing the Codex converter again:
|
||||
|
||||
1. Verify whether the target surface is a skill, a prompt, or both
|
||||
2. Check how Codex derives names from installed artifacts
|
||||
3. Decide which names are canonical before copying deprecated aliases
|
||||
4. Add tests for copied skill content, not just generated prompt content
|
||||
|
||||
## Related Files
|
||||
|
||||
- `src/converters/claude-to-codex.ts`
|
||||
- `src/targets/codex.ts`
|
||||
- `src/types/codex.ts`
|
||||
- `tests/codex-converter.test.ts`
|
||||
- `tests/codex-writer.test.ts`
|
||||
- `README.md`
|
||||
- `plugins/compound-engineering/skills/ce-brainstorm/SKILL.md`
|
||||
- `plugins/compound-engineering/skills/ce-plan/SKILL.md`
|
||||
- `docs/solutions/adding-converter-target-providers.md`
|
||||
@@ -3,6 +3,7 @@ title: Plugin Versioning and Documentation Requirements
|
||||
category: workflow
|
||||
tags: [versioning, changelog, readme, plugin, documentation]
|
||||
created: 2025-11-24
|
||||
date: 2026-03-17
|
||||
severity: process
|
||||
component: plugin-development
|
||||
---
|
||||
@@ -13,65 +14,76 @@ component: plugin-development
|
||||
|
||||
When making changes to the compound-engineering plugin, documentation can get out of sync with the actual components (agents, commands, skills). This leads to confusion about what's included in each version and makes it difficult to track changes over time.
|
||||
|
||||
This document applies to release-owned plugin metadata and changelog surfaces for the `compound-engineering` plugin, not ordinary feature work.
|
||||
|
||||
The broader repo-level release model now lives in:
|
||||
|
||||
- `docs/solutions/workflow/manual-release-please-github-releases.md`
|
||||
|
||||
That doc covers the standing release PR, component ownership across `cli`, `compound-engineering`, `coding-tutor`, and `marketplace`, and the GitHub Releases model for published release notes. This document stays narrower: it is the plugin-scoped reminder for contributors changing `plugins/compound-engineering/**`.
|
||||
|
||||
## Solution
|
||||
|
||||
**Every change to the plugin MUST include:**
|
||||
**Routine PRs should not cut plugin releases.**
|
||||
|
||||
1. **Version bump in `plugin.json`**
|
||||
- Follow semantic versioning (semver)
|
||||
- MAJOR: Breaking changes or major reorganization
|
||||
- MINOR: New agents, commands, or skills added
|
||||
- PATCH: Bug fixes, documentation updates, minor improvements
|
||||
Embedded plugin versions are release-owned metadata. Release automation prepares the next versions and changelog entries after deciding which merged changes ship together. Because multiple PRs may merge before release, contributors should not guess release versions inside individual PRs.
|
||||
|
||||
2. **CHANGELOG.md update**
|
||||
- Add entry under `## [Unreleased]` or new version section
|
||||
- Use Keep a Changelog format
|
||||
- Categories: Added, Changed, Deprecated, Removed, Fixed, Security
|
||||
Contributors should:
|
||||
|
||||
3. **README.md verification**
|
||||
1. **Avoid release bookkeeping in normal PRs**
|
||||
- Do not manually bump `plugins/compound-engineering/.claude-plugin/plugin.json`
|
||||
- Do not manually bump the `compound-engineering` entry in `.claude-plugin/marketplace.json`
|
||||
- Do not cut release sections in the root `CHANGELOG.md`
|
||||
|
||||
2. **Keep substantive docs accurate**
|
||||
- Verify component counts match actual files
|
||||
- Verify agent/command/skill tables are accurate
|
||||
- Update descriptions if functionality changed
|
||||
- Run `bun run release:validate` when plugin inventories or release-owned descriptions may have changed
|
||||
|
||||
## Checklist for Plugin Changes
|
||||
|
||||
```markdown
|
||||
Before committing changes to compound-engineering plugin:
|
||||
|
||||
- [ ] Version bumped in `.claude-plugin/plugin.json`
|
||||
- [ ] CHANGELOG.md updated with changes
|
||||
- [ ] No manual version bump in `plugins/compound-engineering/.claude-plugin/plugin.json`
|
||||
- [ ] No manual version bump in the `compound-engineering` entry inside `.claude-plugin/marketplace.json`
|
||||
- [ ] No manual release section added to `CHANGELOG.md`
|
||||
- [ ] README.md component counts verified
|
||||
- [ ] README.md tables updated (if adding/removing/renaming)
|
||||
- [ ] plugin.json description updated (if component counts changed)
|
||||
- [ ] `bun run release:validate` passes
|
||||
```
|
||||
|
||||
## File Locations
|
||||
|
||||
- Version: `.claude-plugin/plugin.json` → `"version": "X.Y.Z"`
|
||||
- Changelog: `CHANGELOG.md`
|
||||
- Readme: `README.md`
|
||||
- Plugin version is release-owned: `plugins/compound-engineering/.claude-plugin/plugin.json`
|
||||
- Marketplace entry is release-owned: `.claude-plugin/marketplace.json`
|
||||
- Release notes are release-owned: GitHub release PRs and GitHub Releases
|
||||
- Readme: `plugins/compound-engineering/README.md`
|
||||
|
||||
## Example Workflow
|
||||
|
||||
When adding a new agent:
|
||||
|
||||
1. Create the agent file in `agents/[category]/`
|
||||
2. Bump version in `plugin.json` (minor version for new agent)
|
||||
3. Add to CHANGELOG under `### Added`
|
||||
4. Add row to README agent table
|
||||
5. Update README component count
|
||||
6. Update plugin.json description with new counts
|
||||
1. Create the agent file in `plugins/compound-engineering/agents/[category]/`
|
||||
2. Update `plugins/compound-engineering/README.md`
|
||||
3. Leave plugin version selection and canonical release-note generation to release automation
|
||||
4. Run `bun run release:validate`
|
||||
|
||||
## Prevention
|
||||
|
||||
This documentation serves as a reminder. When Claude Code works on this plugin, it should:
|
||||
This documentation serves as a reminder. When maintainers or agents work on this plugin, they should:
|
||||
|
||||
1. Check this doc before committing changes
|
||||
2. Follow the checklist above
|
||||
3. Never commit partial updates (all three files must be updated together)
|
||||
3. Do not guess release versions in feature PRs
|
||||
4. Refer to the repo-level release learning when the question is about batching, release PR behavior, or multi-component ownership rather than plugin-only bookkeeping
|
||||
|
||||
## Related Files
|
||||
|
||||
- `/Users/kieranklaassen/every-marketplace/plugins/compound-engineering/.claude-plugin/plugin.json`
|
||||
- `/Users/kieranklaassen/every-marketplace/plugins/compound-engineering/CHANGELOG.md`
|
||||
- `/Users/kieranklaassen/every-marketplace/plugins/compound-engineering/README.md`
|
||||
- `plugins/compound-engineering/.claude-plugin/plugin.json`
|
||||
- `plugins/compound-engineering/README.md`
|
||||
- `package.json`
|
||||
- `CHANGELOG.md`
|
||||
- `docs/solutions/workflow/manual-release-please-github-releases.md`
|
||||
|
||||
96
docs/solutions/skill-design/beta-skills-framework.md
Normal file
96
docs/solutions/skill-design/beta-skills-framework.md
Normal file
@@ -0,0 +1,96 @@
|
||||
---
|
||||
title: "Beta skills framework: parallel skills with -beta suffix for safe rollouts"
|
||||
category: skill-design
|
||||
date: 2026-03-17
|
||||
module: plugins/compound-engineering/skills
|
||||
component: SKILL.md
|
||||
tags:
|
||||
- skill-design
|
||||
- beta-testing
|
||||
- skill-versioning
|
||||
- rollout-safety
|
||||
severity: medium
|
||||
description: "Pattern for trialing new skill versions alongside stable ones using a -beta suffix. Covers naming, plan file naming, internal references, and promotion path."
|
||||
related:
|
||||
- docs/solutions/skill-design/compound-refresh-skill-improvements.md
|
||||
---
|
||||
|
||||
## Problem
|
||||
|
||||
Core workflow skills like `ce:plan` and `deepen-plan` are deeply chained (`ce:brainstorm` → `ce:plan` → `deepen-plan` → `ce:work`) and orchestrated by `lfg` and `slfg`. Rewriting these skills risks breaking the entire workflow for all users simultaneously. There was no mechanism to let users trial new skill versions alongside stable ones.
|
||||
|
||||
Alternatives considered and rejected:
|
||||
- **Beta gate in SKILL.md** with config-driven routing (`beta: true` in `compound-engineering.local.md`): relies on prompt-level conditional routing which risks instruction blending, requires setup integration, and adds complexity to the skill files themselves.
|
||||
- **Pure router SKILL.md** with both versions in `references/`: adds file-read penalty and refactors stable skills unnecessarily.
|
||||
- **Separate beta plugin**: heavy infrastructure for a temporary need.
|
||||
|
||||
## Solution
|
||||
|
||||
### Parallel skills with `-beta` suffix
|
||||
|
||||
Create separate skill directories alongside the stable ones. Each beta skill is a fully independent copy with its own frontmatter, instructions, and internal references.
|
||||
|
||||
```
|
||||
skills/
|
||||
├── ce-plan/SKILL.md # Stable (unchanged)
|
||||
├── ce-plan-beta/SKILL.md # New version
|
||||
├── deepen-plan/SKILL.md # Stable (unchanged)
|
||||
└── deepen-plan-beta/SKILL.md # New version
|
||||
```
|
||||
|
||||
### Naming and frontmatter conventions
|
||||
|
||||
- **Directory**: `<skill-name>-beta/`
|
||||
- **Frontmatter name**: `<skill:name>-beta` (e.g., `ce:plan-beta`)
|
||||
- **Description**: Write the intended stable description, then prefix with `[BETA]`. This ensures promotion is a simple prefix removal rather than a rewrite.
|
||||
- **`disable-model-invocation: true`**: Prevents the model from auto-triggering the beta skill. Users invoke it manually with the slash command. Remove this field when promoting to stable.
|
||||
- **Plan files**: Use `-beta-plan.md` suffix (e.g., `2026-03-17-001-feat-auth-flow-beta-plan.md`) to avoid clobbering stable plan files
|
||||
|
||||
### Internal references
|
||||
|
||||
Beta skills must reference each other by their beta names:
|
||||
- `ce:plan-beta` references `/deepen-plan-beta` (not `/deepen-plan`)
|
||||
- `deepen-plan-beta` references `ce:plan-beta` (not `ce:plan`)
|
||||
|
||||
### What doesn't change
|
||||
|
||||
- Stable `ce:plan` and `deepen-plan` are completely untouched
|
||||
- `lfg`/`slfg` orchestration continues to use stable skills — no modification needed
|
||||
- `ce:brainstorm` still hands off to stable `ce:plan` — no modification needed
|
||||
- `ce:work` consumes plan files from either version (reads the file, doesn't care which skill wrote it)
|
||||
|
||||
### Tradeoffs
|
||||
|
||||
**Simplicity over seamless integration.** Beta skills exist as standalone, manually-invoked skills. They won't be auto-triggered by `ce:brainstorm` handoffs or `lfg`/`slfg` orchestration without further surgery to those skills, which isn't worth the complexity for a trial period.
|
||||
|
||||
**Intended usage pattern:** A user can run `/ce:plan` for the stable output, then run `/ce:plan-beta` on the same input to compare the two plan documents side by side. The `-beta-plan.md` suffix ensures both outputs coexist in `docs/plans/` without collision.
|
||||
|
||||
## Promotion path
|
||||
|
||||
When the beta version is validated:
|
||||
|
||||
1. Replace stable `SKILL.md` content with beta skill content
|
||||
2. Restore stable frontmatter: remove `[BETA]` prefix from description, restore stable `name:`
|
||||
3. Remove `disable-model-invocation: true` so the model can auto-trigger it
|
||||
4. Update all internal references back to stable names
|
||||
5. Restore stable plan file naming (remove `-beta` from the convention)
|
||||
6. Delete the beta skill directory
|
||||
7. Update README.md: remove from Beta Skills section, verify counts
|
||||
8. Verify `lfg`/`slfg` work with the promoted skill
|
||||
9. Verify `ce:work` consumes plans from the promoted skill
|
||||
|
||||
## Validation
|
||||
|
||||
After creating a beta skill, search its SKILL.md for references to the stable skill name it replaces. Any occurrence of the stable name without `-beta` is a missed rename — it would cause output collisions or route to the wrong skill.
|
||||
|
||||
Check for:
|
||||
- **Output file paths** that use the stable naming convention instead of the `-beta` variant
|
||||
- **Cross-skill references** that point to stable skill names instead of beta counterparts
|
||||
- **User-facing text** (questions, confirmations) that mentions stable paths or names
|
||||
|
||||
## Prevention
|
||||
|
||||
- When adding a beta skill, always use the `-beta` suffix consistently in directory name, frontmatter name, description, plan file naming, and all internal skill-to-skill references
|
||||
- After creating a beta skill, run the validation checks above to catch missed renames in file paths, user-facing text, and cross-skill references
|
||||
- Always test that stable skills are completely unaffected by the beta skill's existence
|
||||
- Keep beta and stable plan file suffixes distinct so outputs can coexist for comparison
|
||||
@@ -0,0 +1,141 @@
|
||||
---
|
||||
title: "ce:compound-refresh skill redesign for autonomous maintenance without live user context"
|
||||
category: skill-design
|
||||
date: 2026-03-13
|
||||
module: plugins/compound-engineering/skills/ce-compound-refresh
|
||||
component: SKILL.md
|
||||
tags:
|
||||
- skill-design
|
||||
- compound-refresh
|
||||
- maintenance-workflow
|
||||
- drift-classification
|
||||
- subagent-architecture
|
||||
- platform-agnostic
|
||||
severity: medium
|
||||
description: "Redesign ce:compound-refresh to handle autonomous drift triage, in-skill replacement via subagents, and smart scoping without relying on live problem-solving context that ce:compound expects."
|
||||
related:
|
||||
- docs/solutions/plugin-versioning-requirements.md
|
||||
- https://github.com/EveryInc/compound-engineering-plugin/pull/260
|
||||
- https://github.com/EveryInc/compound-engineering-plugin/issues/204
|
||||
- https://github.com/EveryInc/compound-engineering-plugin/issues/221
|
||||
---
|
||||
|
||||
## Problem
|
||||
|
||||
The initial `ce:compound-refresh` skill had several design issues discovered during real-world testing:
|
||||
|
||||
1. Interactive questions never triggered the proper tool (AskUserQuestion) because the instruction used a weak "when available" qualifier
|
||||
2. Auto-archive criteria contradicted a "always ask before archiving" rule in a later phase
|
||||
3. Broad scope (9+ docs) asked the user to choose an area blindly without providing analysis
|
||||
4. The Replace flow tried to hand off to `ce:compound`, which expects fresh problem-solving context the user doesn't have months later
|
||||
5. Subagents used shell commands for file existence checks, triggering permission prompts
|
||||
6. No way to run the skill unattended (e.g., on a schedule) — every run required user interaction
|
||||
|
||||
## Root Cause
|
||||
|
||||
Five independent design issues, each with a distinct root cause:
|
||||
|
||||
1. **Hardcoded tool name with escape hatch.** Saying "Use AskUserQuestion when available" gave the model permission to skip the tool and just output text. Also non-portable to Codex and other platforms.
|
||||
2. **Contradictory rules across phases.** Phase 2 defined auto-archive criteria. Phase 3 said "always ask before archiving" with no exception. The model followed Phase 3.
|
||||
3. **Question before evidence.** The skill prompted scope selection before gathering any information about which areas were most stale or interconnected.
|
||||
4. **Unsatisfied precondition in cross-skill handoff.** `ce:compound` expects a recently solved problem with fresh context. A maintenance refresh has investigation evidence instead — equivalent data, different shape.
|
||||
5. **No tool preference guidance for subagents.** Without explicit instruction, subagents defaulted to bash for file operations.
|
||||
6. **Interactive-only design.** Every phase assumed a user was present. No way to run autonomously for scheduled maintenance or hands-off sweeps.
|
||||
|
||||
## Solution
|
||||
|
||||
### 1. Platform-agnostic interactive questions
|
||||
|
||||
Reference "the platform's interactive question tool" as the concept, with concrete examples:
|
||||
|
||||
```markdown
|
||||
Ask questions **one at a time** — use the platform's interactive question tool
|
||||
(e.g. `AskUserQuestion` in Claude Code, `request_user_input` in Codex) and
|
||||
**stop to wait for the answer** before continuing.
|
||||
```
|
||||
|
||||
The "stop to wait" language removes the escape hatch. The examples help each platform's model select the right tool.
|
||||
|
||||
### 2. Auto-archive exemption for unambiguous cases
|
||||
|
||||
Phase 3 now defers to Phase 2's auto-archive criteria:
|
||||
|
||||
```markdown
|
||||
You are about to Archive a document **and** the evidence is not unambiguous
|
||||
(see auto-archive criteria in Phase 2). When auto-archive criteria are met,
|
||||
proceed without asking.
|
||||
```
|
||||
|
||||
### 3. Smart triage for broad scope
|
||||
|
||||
When 9+ candidate docs are found, triage before asking:
|
||||
|
||||
1. **Inventory** — read frontmatter, group by module/component/category
|
||||
2. **Impact clustering** — dense clusters of interconnected learnings + pattern docs are higher-impact than isolated docs
|
||||
3. **Spot-check drift** — check whether primary referenced files still exist
|
||||
4. **Recommend** — present the highest-impact cluster with rationale
|
||||
|
||||
Key insight: "code changed recently" is NOT a reliable staleness signal. Missing references in a high-impact cluster is the strongest signal.
|
||||
|
||||
### 4. Replacement subagents instead of ce:compound handoff
|
||||
|
||||
By the time a Replace is identified, Phase 1 investigation has already gathered the evidence that `ce:compound` would research:
|
||||
- The old learning's claims
|
||||
- What the current code actually does
|
||||
- Where and why the drift occurred
|
||||
|
||||
A replacement subagent writes the successor directly using `ce:compound`'s document format (frontmatter, problem, root cause, solution, prevention). Run sequentially — one at a time — because each may read significant code.
|
||||
|
||||
When evidence is insufficient (e.g., entire subsystem replaced, new architecture too complex to understand from investigation alone), mark as stale and recommend `ce:compound` after the user's next encounter with that area.
|
||||
|
||||
### 5. Dedicated file tools over shell commands
|
||||
|
||||
Added to subagent strategy:
|
||||
|
||||
```markdown
|
||||
Subagents should use dedicated file search and read tools for investigation —
|
||||
not shell commands. This avoids unnecessary permission prompts and is more
|
||||
reliable across platforms.
|
||||
```
|
||||
|
||||
### 6. Autonomous mode for scheduled/unattended runs
|
||||
|
||||
Added `mode:autonomous` argument support so the skill can run without user interaction (e.g., on a schedule, in CI, or when the user just wants a hands-off sweep).
|
||||
|
||||
Key design decisions:
|
||||
- **Explicit opt-in only.** `mode:autonomous` must be in the arguments. Auto-detection based on tool availability was rejected because a user in an interactive agent without a question tool (e.g., Cursor, Windsurf) is still interactive — they just use plain-text replies.
|
||||
- **Conservative confidence.** Borderline cases that would get a user question in interactive mode get marked stale in autonomous mode. Err toward stale-marking over incorrect action.
|
||||
- **Detailed report as deliverable.** Since no user was present, the output report includes full rationale for each action so a human can review after the fact.
|
||||
- **Process everything.** No scope narrowing questions — if no scope hint provided, process all docs. For broad scope, process clusters in impact order without asking.
|
||||
|
||||
## Prevention
|
||||
|
||||
### Skill review checklist additions
|
||||
|
||||
These five patterns should be checked during any skill review:
|
||||
|
||||
1. **No hardcoded tool names** — All tool references use capability-first language with platform examples and a plain-text fallback
|
||||
2. **No contradictory rules across phases** — Trace each action type through all phases; verify absolute language ("always," "never") is not contradicted elsewhere
|
||||
3. **No blind user questions** — Every question presented to the user is informed by evidence the agent gathered first
|
||||
4. **No unsatisfied cross-skill preconditions** — Every skill handoff verifies the target skill's preconditions are met by the calling context
|
||||
5. **No shell commands for file operations in subagents** — Subagent instructions explicitly prefer dedicated tools over shell commands
|
||||
6. **Autonomous mode for long-running skills** — Any skill that could run unattended should support an explicit opt-in mode with conservative confidence and detailed reporting
|
||||
|
||||
### Key anti-patterns
|
||||
|
||||
| Anti-pattern | Better pattern |
|
||||
|---|---|
|
||||
| "Use the AskUserQuestion tool when available" | "Use the platform's interactive question tool (e.g. AskUserQuestion in Claude Code, request_user_input in Codex)" |
|
||||
| Defining auto-archive conditions, then "always ask before archiving" | Single-source-of-truth: define the rule once, reference it elsewhere |
|
||||
| "Which area should we review?" before any investigation | Triage first, recommend with evidence, let user confirm or redirect |
|
||||
| "Create a successor learning through ce:compound" during a refresh | Replacement subagent writes directly using gathered evidence |
|
||||
| No tool guidance for subagents | "Use dedicated file search and read tools, not shell commands" |
|
||||
| Auto-detecting "no question tool = headless" | Explicit `mode:autonomous` argument — interactive agents without question tools are still interactive |
|
||||
|
||||
## Cross-References
|
||||
|
||||
- **PR #260**: The PR containing all these improvements
|
||||
- **Issue #204**: Platform-agnostic tool references (AskUserQuestion dependency)
|
||||
- **Issue #221**: Motivating issue for maintenance at scale
|
||||
- **PR #242**: ce:audit (detection counterpart, closed)
|
||||
- **PR #150**: Established subagent context-isolation pattern
|
||||
210
docs/solutions/workflow/manual-release-please-github-releases.md
Normal file
210
docs/solutions/workflow/manual-release-please-github-releases.md
Normal file
@@ -0,0 +1,210 @@
|
||||
---
|
||||
title: "Manual release-please with GitHub Releases for multi-component plugin and marketplace releases"
|
||||
category: workflow
|
||||
date: 2026-03-17
|
||||
created: 2026-03-17
|
||||
severity: process
|
||||
component: release-automation
|
||||
tags:
|
||||
- release-please
|
||||
- semantic-release
|
||||
- github-releases
|
||||
- marketplace
|
||||
- plugin-versioning
|
||||
- ci
|
||||
- automation
|
||||
- release-process
|
||||
---
|
||||
|
||||
# Manual release-please with GitHub Releases for multi-component plugin and marketplace releases
|
||||
|
||||
## Problem
|
||||
|
||||
The repo had one automated release path for the npm CLI, but the actual release model was fragmented across:
|
||||
|
||||
- root-only `semantic-release`
|
||||
- a local maintainer workflow via `release-docs`
|
||||
- multiple version-bearing metadata files
|
||||
- inconsistent release-note ownership
|
||||
|
||||
That made it hard to batch merges on `main`, hard for multiple maintainers to share release responsibility, and easy for release notes, plugin manifests, marketplace metadata, and computed counts to drift out of sync.
|
||||
|
||||
## Root Cause
|
||||
|
||||
Release intent, component ownership, release-note ownership, and metadata synchronization were split across different systems:
|
||||
|
||||
- PRs merged to `main` were too close to an actual publish event
|
||||
- only the root CLI had a real CI-owned release path
|
||||
- plugin and marketplace releases depended on local knowledge and stale docs
|
||||
- the repo had multiple release surfaces (`cli`, `compound-engineering`, `coding-tutor`, `marketplace`) but no single release authority
|
||||
|
||||
An adjacent contributor-guidance problem made this worse: root `CLAUDE.md` had become a large, stale, partially duplicated instruction file, while `AGENTS.md` was the better canonical repo guidance surface.
|
||||
|
||||
## Solution
|
||||
|
||||
Move the repo to a manual `release-please` model with one standing release PR and explicit component ownership.
|
||||
|
||||
Key decisions:
|
||||
|
||||
- Use `release-please` manifest mode for four release components:
|
||||
- `cli`
|
||||
- `compound-engineering`
|
||||
- `coding-tutor`
|
||||
- `marketplace`
|
||||
- Keep release timing manual: the actual release happens when the generated release PR is merged.
|
||||
- Keep release PR maintenance automatic on pushes to `main`.
|
||||
- Use GitHub release PRs and GitHub Releases as the canonical release-notes surface for new releases.
|
||||
- Replace `release-docs` with repo-owned scripts for preview, metadata sync, and validation.
|
||||
- Keep PR title scopes optional; use file paths to determine affected components.
|
||||
- Make `AGENTS.md` canonical and reduce root `CLAUDE.md` to a compatibility shim.
|
||||
|
||||
## Critical Constraint Discovered
|
||||
|
||||
`release-please` does not allow package changelog paths that traverse upward with `..`.
|
||||
|
||||
The failed first live run exposed this directly:
|
||||
|
||||
- `release-please failed: illegal pathing characters in path: plugins/compound-engineering/../../CHANGELOG.md`
|
||||
|
||||
That means a multi-component repo cannot force subpackage release entries back into one shared root changelog file using `changelog-path` values like:
|
||||
|
||||
- `../../CHANGELOG.md`
|
||||
- `../CHANGELOG.md`
|
||||
|
||||
The practical fix was:
|
||||
|
||||
- set `skip-changelog: true` for all components in `.github/release-please-config.json`
|
||||
- treat GitHub Releases as the canonical release-notes surface
|
||||
- reduce `CHANGELOG.md` to a simple pointer file
|
||||
- add repo validation to catch illegal upward changelog paths before merge
|
||||
|
||||
## Resulting Release Process
|
||||
|
||||
After the migration:
|
||||
|
||||
1. Normal feature PRs merge to `main`.
|
||||
2. The `Release PR` workflow updates one standing release PR for the repo.
|
||||
3. Additional releasable merges accumulate into that same release PR.
|
||||
4. Maintainers can inspect the standing release PR or run the manual preview flow.
|
||||
5. The actual release happens only when the generated release PR is merged.
|
||||
6. npm publish runs only when the `cli` component is part of that release.
|
||||
7. Component-specific release notes are published via GitHub releases such as `cli-vX.Y.Z` and `compound-engineering-vX.Y.Z`.
|
||||
|
||||
## Component Rules
|
||||
|
||||
- PR title determines release intent:
|
||||
- `feat` => minor
|
||||
- `fix` / `perf` / `refactor` / `revert` => patch
|
||||
- `!` => major
|
||||
- File paths determine component ownership:
|
||||
- `src/**`, `package.json`, `bun.lock`, `tests/cli.test.ts` => `cli`
|
||||
- `plugins/compound-engineering/**` => `compound-engineering`
|
||||
- `plugins/coding-tutor/**` => `coding-tutor`
|
||||
- `.claude-plugin/marketplace.json` => `marketplace`
|
||||
- Optional title scopes are advisory only.
|
||||
|
||||
This keeps titles simple while still letting the release system decide the correct component bump.
|
||||
|
||||
## Examples
|
||||
|
||||
### One merge lands, but no release is cut yet
|
||||
|
||||
- A `fix:` PR merges to `main`
|
||||
- The standing release PR updates
|
||||
- Nothing is published yet
|
||||
|
||||
### More work lands before release
|
||||
|
||||
- A later `feat:` PR merges to `main`
|
||||
- The same open release PR updates to include both changes
|
||||
- The pending bump can increase based on total unreleased work
|
||||
|
||||
### Plugin-only release
|
||||
|
||||
- A change lands only under `plugins/coding-tutor/**`
|
||||
- Only `coding-tutor` should bump
|
||||
- `compound-engineering`, `marketplace`, and `cli` should remain untouched
|
||||
- npm publish should not run unless `cli` is also part of that release
|
||||
|
||||
### Marketplace-only release
|
||||
|
||||
- A new plugin is added to the catalog or marketplace metadata changes
|
||||
- `marketplace` bumps
|
||||
- Existing plugin versions do not need to bump just because the catalog changed
|
||||
|
||||
### Exceptional manual bump
|
||||
|
||||
- Maintainers decide the inferred bump is too small
|
||||
- They use the preview/release override path instead of making fake commits
|
||||
- The release still goes through the same CI-owned process
|
||||
|
||||
## Release Notes Model
|
||||
|
||||
- Pending release state is visible in one standing release PR.
|
||||
- Published release history is canonical in GitHub Releases.
|
||||
- Component identity is carried by component-specific tags such as:
|
||||
- `cli-vX.Y.Z`
|
||||
- `compound-engineering-vX.Y.Z`
|
||||
- `coding-tutor-vX.Y.Z`
|
||||
- `marketplace-vX.Y.Z`
|
||||
- Root `CHANGELOG.md` is only a pointer to GitHub Releases and is not the canonical source for new releases.
|
||||
|
||||
## Key Files
|
||||
|
||||
- `.github/release-please-config.json`
|
||||
- `.github/.release-please-manifest.json`
|
||||
- `.github/workflows/release-pr.yml`
|
||||
- `.github/workflows/release-preview.yml`
|
||||
- `.github/workflows/ci.yml`
|
||||
- `src/release/components.ts`
|
||||
- `src/release/metadata.ts`
|
||||
- `scripts/release/preview.ts`
|
||||
- `scripts/release/sync-metadata.ts`
|
||||
- `scripts/release/validate.ts`
|
||||
- `AGENTS.md`
|
||||
- `CLAUDE.md`
|
||||
|
||||
## Prevention
|
||||
|
||||
- Keep release authority in CI only.
|
||||
- Do not reintroduce local maintainer-only release flows or hand-managed version bumps.
|
||||
- Keep `AGENTS.md` canonical. If a tool still needs `CLAUDE.md`, use it only as a compatibility shim.
|
||||
- Do not try to force multi-component release notes back into one committed changelog file if the tool does not support it natively.
|
||||
- Validate `.github/release-please-config.json` in CI so unsupported changelog-path values fail before the workflow reaches GitHub Actions.
|
||||
- Run `bun run release:validate` whenever plugin inventories, release-owned descriptions, or marketplace entries may have changed.
|
||||
- Prefer maintained CI actions over custom validation when a generic concern does not need repo-specific logic.
|
||||
|
||||
## Validation Checklist
|
||||
|
||||
Before merge:
|
||||
|
||||
- Confirm PR title passes semantic validation.
|
||||
- Run `bun test`.
|
||||
- Run `bun run release:validate`.
|
||||
- Run `bun run release:preview ...` for representative changed files.
|
||||
|
||||
After merging release-system changes to `main`:
|
||||
|
||||
- Verify exactly one standing release PR is created or updated.
|
||||
- Confirm ordinary merges to `main` do not publish npm directly.
|
||||
- Inspect the release PR for correct component selection, versions, and metadata updates.
|
||||
|
||||
Before merging a generated release PR:
|
||||
|
||||
- Verify untouched components are unchanged.
|
||||
- Verify `marketplace` only bumps for marketplace-level changes.
|
||||
- Verify plugin-only changes do not imply `cli` unless `src/` also changed.
|
||||
|
||||
After merging a generated release PR:
|
||||
|
||||
- Confirm npm publish runs only when `cli` is part of the release.
|
||||
- Confirm no recursive follow-up release PR appears containing only generated churn.
|
||||
- Confirm the expected component GitHub releases were created and that release-owned metadata matches the released components.
|
||||
|
||||
## Related Docs
|
||||
|
||||
- `docs/solutions/plugin-versioning-requirements.md`
|
||||
- `docs/solutions/adding-converter-target-providers.md`
|
||||
- `AGENTS.md`
|
||||
- `plugins/compound-engineering/AGENTS.md`
|
||||
- `docs/specs/kiro.md`
|
||||
@@ -48,7 +48,9 @@ https://developers.openai.com/codex/mcp
|
||||
- `SKILL.md` uses YAML front matter and requires `name` and `description`. citeturn3view3turn3view4
|
||||
- Required fields are single-line with length limits (name ≤ 100 chars, description ≤ 500 chars). citeturn3view4
|
||||
- At startup, Codex loads only each skill’s name/description; full content is injected when invoked. citeturn3view3turn3view4
|
||||
- Skills can be repo-scoped in `.codex/skills/` or user-scoped in `~/.codex/skills/`. citeturn3view4
|
||||
- Skills can be repo-scoped in `.agents/skills/` and are discovered from the current working directory up to the repository root. User-scoped skills live in `~/.agents/skills/`. citeturn1view1turn1view4
|
||||
- Inference: some existing tooling and user setups still use `.codex/skills/` and `~/.codex/skills/` as legacy compatibility paths, but those locations are not documented in the current OpenAI Codex skills docs linked above.
|
||||
- Codex also supports admin-scoped skills in `/etc/codex/skills` plus built-in system skills bundled with Codex. citeturn1view4
|
||||
- Skills can be invoked explicitly using `/skills` or `$skill-name`. citeturn3view3
|
||||
|
||||
## MCP (Model Context Protocol)
|
||||
|
||||
122
docs/specs/copilot.md
Normal file
122
docs/specs/copilot.md
Normal file
@@ -0,0 +1,122 @@
|
||||
# GitHub Copilot Spec (Agents, Skills, MCP)
|
||||
|
||||
Last verified: 2026-02-14
|
||||
|
||||
## Primary sources
|
||||
|
||||
```
|
||||
https://docs.github.com/en/copilot/reference/custom-agents-configuration
|
||||
https://docs.github.com/en/copilot/concepts/agents/about-agent-skills
|
||||
https://docs.github.com/en/copilot/concepts/agents/coding-agent/mcp-and-coding-agent
|
||||
```
|
||||
|
||||
## Config locations
|
||||
|
||||
| Scope | Path |
|
||||
|-------|------|
|
||||
| Project agents | `.github/agents/*.agent.md` |
|
||||
| Project skills | `.github/skills/*/SKILL.md` |
|
||||
| Project instructions | `.github/copilot-instructions.md` |
|
||||
| Path-specific instructions | `.github/instructions/*.instructions.md` |
|
||||
| Project prompts | `.github/prompts/*.prompt.md` |
|
||||
| Org/enterprise agents | `.github-private/agents/*.agent.md` |
|
||||
| Personal skills | `~/.copilot/skills/*/SKILL.md` |
|
||||
| Directory instructions | `AGENTS.md` (nearest ancestor wins) |
|
||||
|
||||
## Agents (.agent.md files)
|
||||
|
||||
- Custom agents are Markdown files with YAML frontmatter stored in `.github/agents/`.
|
||||
- File extension is `.agent.md` (or `.md`). Filenames may only contain: `.`, `-`, `_`, `a-z`, `A-Z`, `0-9`.
|
||||
- `description` is the only required frontmatter field.
|
||||
|
||||
### Frontmatter fields
|
||||
|
||||
| Field | Required | Default | Description |
|
||||
|-------|----------|---------|-------------|
|
||||
| `name` | No | Derived from filename | Display name |
|
||||
| `description` | **Yes** | — | What the agent does |
|
||||
| `tools` | No | `["*"]` | Tool access list. `[]` disables all tools. |
|
||||
| `target` | No | both | `vscode`, `github-copilot`, or omit for both |
|
||||
| `infer` | No | `true` | Auto-select based on task context |
|
||||
| `model` | No | Platform default | AI model (works in IDE, may be ignored on github.com) |
|
||||
| `mcp-servers` | No | — | MCP config (org/enterprise agents only) |
|
||||
| `metadata` | No | — | Arbitrary key-value annotations |
|
||||
|
||||
### Character limit
|
||||
|
||||
Agent body content is limited to **30,000 characters**.
|
||||
|
||||
### Tool names
|
||||
|
||||
| Name | Aliases | Purpose |
|
||||
|------|---------|---------|
|
||||
| `execute` | `shell`, `Bash` | Run shell commands |
|
||||
| `read` | `Read` | Read files |
|
||||
| `edit` | `Edit`, `Write` | Modify files |
|
||||
| `search` | `Grep`, `Glob` | Search files |
|
||||
| `agent` | `Task` | Invoke other agents |
|
||||
| `web` | `WebSearch`, `WebFetch` | Web access |
|
||||
|
||||
## Skills (SKILL.md)
|
||||
|
||||
- Skills follow the open SKILL.md standard (same format as Claude Code and Cursor).
|
||||
- A skill is a directory containing `SKILL.md` plus optional `scripts/`, `references/`, and `assets/`.
|
||||
- YAML frontmatter requires `name` and `description` fields.
|
||||
- Skills are loaded on-demand when Copilot determines relevance.
|
||||
|
||||
### Discovery locations
|
||||
|
||||
| Scope | Path |
|
||||
|-------|------|
|
||||
| Project | `.github/skills/*/SKILL.md` |
|
||||
| Project (Claude-compatible) | `.claude/skills/*/SKILL.md` |
|
||||
| Project (auto-discovery) | `.agents/skills/*/SKILL.md` |
|
||||
| Personal | `~/.copilot/skills/*/SKILL.md` |
|
||||
|
||||
## MCP (Model Context Protocol)
|
||||
|
||||
- MCP configuration is set via **Repository Settings > Copilot > Coding agent > MCP configuration** on GitHub.
|
||||
- Repository-level agents **cannot** define MCP servers inline; use repository settings instead.
|
||||
- Org/enterprise agents can embed MCP server definitions in frontmatter.
|
||||
- All env var names must use the `COPILOT_MCP_` prefix.
|
||||
- Only MCP tools are supported (not resources or prompts).
|
||||
|
||||
### Config structure
|
||||
|
||||
```json
|
||||
{
|
||||
"mcpServers": {
|
||||
"server-name": {
|
||||
"type": "local",
|
||||
"command": "npx",
|
||||
"args": ["package"],
|
||||
"tools": ["*"],
|
||||
"env": {
|
||||
"API_KEY": "COPILOT_MCP_API_KEY"
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
### Server types
|
||||
|
||||
| Type | Fields |
|
||||
|------|--------|
|
||||
| Local/stdio | `type: "local"`, `command`, `args`, `tools`, `env` |
|
||||
| Remote/SSE | `type: "sse"`, `url`, `tools`, `headers` |
|
||||
|
||||
## Prompts (.prompt.md)
|
||||
|
||||
- Reusable prompt files stored in `.github/prompts/`.
|
||||
- Available in VS Code, Visual Studio, and JetBrains IDEs only (not on github.com).
|
||||
- Invoked via `/promptname` in chat.
|
||||
- Support variable syntax: `${input:name}`, `${file}`, `${selection}`.
|
||||
|
||||
## Precedence
|
||||
|
||||
1. Repository-level agents
|
||||
2. Organization-level agents (`.github-private`)
|
||||
3. Enterprise-level agents (`.github-private`)
|
||||
|
||||
Within a repo, `AGENTS.md` files in directories provide nearest-ancestor-wins instructions.
|
||||
171
docs/specs/kiro.md
Normal file
171
docs/specs/kiro.md
Normal file
@@ -0,0 +1,171 @@
|
||||
# Kiro CLI Spec (Custom Agents, Skills, Steering, MCP, Settings)
|
||||
|
||||
Last verified: 2026-02-17
|
||||
|
||||
## Primary sources
|
||||
|
||||
```
|
||||
https://kiro.dev/docs/cli/
|
||||
https://kiro.dev/docs/cli/custom-agents/configuration-reference/
|
||||
https://kiro.dev/docs/cli/skills/
|
||||
https://kiro.dev/docs/cli/steering/
|
||||
https://kiro.dev/docs/cli/mcp/
|
||||
https://kiro.dev/docs/cli/hooks/
|
||||
https://agentskills.io
|
||||
```
|
||||
|
||||
## Config locations
|
||||
|
||||
- Project-level config: `.kiro/` directory at project root.
|
||||
- No global/user-level config directory — all config is project-scoped.
|
||||
|
||||
## Directory structure
|
||||
|
||||
```
|
||||
.kiro/
|
||||
├── agents/
|
||||
│ ├── <name>.json # Agent configuration
|
||||
│ └── prompts/
|
||||
│ └── <name>.md # Agent prompt files
|
||||
├── skills/
|
||||
│ └── <name>/
|
||||
│ └── SKILL.md # Skill definition
|
||||
├── steering/
|
||||
│ └── <name>.md # Always-on context files
|
||||
└── settings/
|
||||
└── mcp.json # MCP server configuration
|
||||
```
|
||||
|
||||
## Custom agents (JSON config + prompt files)
|
||||
|
||||
- Custom agents are JSON files in `.kiro/agents/`.
|
||||
- Each agent has a corresponding prompt `.md` file, referenced via `file://` URI.
|
||||
- Agent config has 14 possible fields (see below).
|
||||
- Agents are activated by user selection (no auto-activation).
|
||||
- The converter outputs a subset of fields relevant to converted plugins.
|
||||
|
||||
### Agent config fields
|
||||
|
||||
| Field | Type | Used in conversion | Notes |
|
||||
|---|---|---|---|
|
||||
| `name` | string | Yes | Agent display name |
|
||||
| `description` | string | Yes | Human-readable description |
|
||||
| `prompt` | string or `file://` URI | Yes | System prompt or file reference |
|
||||
| `tools` | string[] | Yes (`["*"]`) | Available tools |
|
||||
| `resources` | string[] | Yes | `file://`, `skill://`, `knowledgeBase` URIs |
|
||||
| `includeMcpJson` | boolean | Yes (`true`) | Inherit project MCP servers |
|
||||
| `welcomeMessage` | string | Yes | Agent switch greeting |
|
||||
| `mcpServers` | object | No | Per-agent MCP config (use includeMcpJson instead) |
|
||||
| `toolAliases` | Record | No | Tool name remapping |
|
||||
| `allowedTools` | string[] | No | Auto-approve patterns |
|
||||
| `toolsSettings` | object | No | Per-tool configuration |
|
||||
| `hooks` | object | No (future work) | 5 trigger types |
|
||||
| `model` | string | No | Model selection |
|
||||
| `keyboardShortcut` | string | No | Quick-switch shortcut |
|
||||
|
||||
### Example agent config
|
||||
|
||||
```json
|
||||
{
|
||||
"name": "security-reviewer",
|
||||
"description": "Reviews code for security vulnerabilities",
|
||||
"prompt": "file://./prompts/security-reviewer.md",
|
||||
"tools": ["*"],
|
||||
"resources": [
|
||||
"file://.kiro/steering/**/*.md",
|
||||
"skill://.kiro/skills/**/SKILL.md"
|
||||
],
|
||||
"includeMcpJson": true,
|
||||
"welcomeMessage": "Switching to security-reviewer. Reviews code for security vulnerabilities"
|
||||
}
|
||||
```
|
||||
|
||||
## Skills (SKILL.md standard)
|
||||
|
||||
- Skills follow the open [Agent Skills](https://agentskills.io) standard.
|
||||
- A skill is a folder containing `SKILL.md` plus optional supporting files.
|
||||
- Skills live in `.kiro/skills/`.
|
||||
- `SKILL.md` uses YAML frontmatter with `name` and `description` fields.
|
||||
- Kiro activates skills on demand based on description matching.
|
||||
- The `description` field is critical — Kiro uses it to decide when to activate the skill.
|
||||
|
||||
### Constraints
|
||||
|
||||
- Skill name: max 64 characters, pattern `^[a-z][a-z0-9-]*$`, no consecutive hyphens (`--`).
|
||||
- Skill description: max 1024 characters.
|
||||
- Skill name must match parent directory name.
|
||||
|
||||
### Example
|
||||
|
||||
```yaml
|
||||
---
|
||||
name: workflows-plan
|
||||
description: Plan work by analyzing requirements and creating actionable steps
|
||||
---
|
||||
|
||||
# Planning Workflow
|
||||
|
||||
Detailed instructions...
|
||||
```
|
||||
|
||||
## Steering files
|
||||
|
||||
- Markdown files in `.kiro/steering/`.
|
||||
- Always loaded into every agent session's context.
|
||||
- Equivalent to the repo instruction file used by Claude-oriented workflows; in this repo `AGENTS.md` is canonical and `CLAUDE.md` may exist only as a compatibility shim.
|
||||
- Used for project-wide instructions, coding standards, and conventions.
|
||||
|
||||
## MCP server configuration
|
||||
|
||||
- MCP servers are configured in `.kiro/settings/mcp.json`.
|
||||
- **Only stdio transport is supported** — `command` + `args` + `env`.
|
||||
- HTTP/SSE transport (`url`, `headers`) is NOT supported by Kiro CLI.
|
||||
- The converter skips HTTP-only MCP servers with a warning.
|
||||
|
||||
### Example
|
||||
|
||||
```json
|
||||
{
|
||||
"mcpServers": {
|
||||
"playwright": {
|
||||
"command": "npx",
|
||||
"args": ["-y", "@anthropic/mcp-playwright"]
|
||||
},
|
||||
"context7": {
|
||||
"command": "npx",
|
||||
"args": ["-y", "@context7/mcp-server"]
|
||||
}
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
## Hooks
|
||||
|
||||
- Kiro supports 5 hook trigger types: `agentSpawn`, `userPromptSubmit`, `preToolUse`, `postToolUse`, `stop`.
|
||||
- Hooks are configured inside agent JSON configs (not separate files).
|
||||
- 3 of 5 triggers map to Claude Code hooks (`preToolUse`, `postToolUse`, `stop`).
|
||||
- Not converted by the plugin converter for MVP — a warning is emitted.
|
||||
|
||||
## Conversion lossy mappings
|
||||
|
||||
| Claude Code Feature | Kiro Status | Notes |
|
||||
|---|---|---|
|
||||
| `Edit` tool (surgical replacement) | Degraded -> `write` (full-file) | Kiro write overwrites entire files |
|
||||
| `context: fork` | Lost | No execution isolation control |
|
||||
| `!`command`` dynamic injection | Lost | No pre-processing of markdown |
|
||||
| `disable-model-invocation` | Lost | No invocation control |
|
||||
| `allowed-tools` per skill | Lost | No tool permission scoping per skill |
|
||||
| `$ARGUMENTS` interpolation | Lost | No structured argument passing |
|
||||
| Claude hooks | Skipped | Future follow-up (near-1:1 for 3/5 triggers) |
|
||||
| HTTP MCP servers | Skipped | Kiro only supports stdio transport |
|
||||
|
||||
## Overwrite behavior during conversion
|
||||
|
||||
| Content Type | Strategy | Rationale |
|
||||
|---|---|---|
|
||||
| Generated agents (JSON + prompt) | Overwrite | Generated, not user-authored |
|
||||
| Generated skills (from commands) | Overwrite | Generated, not user-authored |
|
||||
| Copied skills (pass-through) | Overwrite | Plugin is source of truth |
|
||||
| Steering files | Overwrite | Generated from `AGENTS.md` when present, otherwise `CLAUDE.md` |
|
||||
| `mcp.json` | Merge with backup | User may have added their own servers |
|
||||
| User-created agents/skills | Preserved | Don't delete orphans |
|
||||
477
docs/specs/windsurf.md
Normal file
477
docs/specs/windsurf.md
Normal file
@@ -0,0 +1,477 @@
|
||||
# Windsurf Editor Global Configuration Guide
|
||||
|
||||
> **Purpose**: Technical reference for programmatically creating and managing Windsurf's global Skills, Workflows, and Rules.
|
||||
>
|
||||
> **Source**: Official Windsurf documentation at [docs.windsurf.com](https://docs.windsurf.com) + local file analysis.
|
||||
>
|
||||
> **Last Updated**: February 2026
|
||||
|
||||
---
|
||||
|
||||
## Table of Contents
|
||||
|
||||
1. [Overview](#overview)
|
||||
2. [Base Directory Structure](#base-directory-structure)
|
||||
3. [Skills](#skills)
|
||||
4. [Workflows](#workflows)
|
||||
5. [Rules](#rules)
|
||||
6. [Memories](#memories)
|
||||
7. [System-Level Configuration (Enterprise)](#system-level-configuration-enterprise)
|
||||
8. [Programmatic Creation Reference](#programmatic-creation-reference)
|
||||
9. [Best Practices](#best-practices)
|
||||
|
||||
---
|
||||
|
||||
## Overview
|
||||
|
||||
Windsurf provides three main customization mechanisms:
|
||||
|
||||
| Feature | Purpose | Invocation |
|
||||
|---------|---------|------------|
|
||||
| **Skills** | Complex multi-step tasks with supporting resources | Automatic (progressive disclosure) or `@skill-name` |
|
||||
| **Workflows** | Reusable step-by-step procedures | Slash command `/workflow-name` |
|
||||
| **Rules** | Behavioral guidelines and preferences | Trigger-based (always-on, glob, manual, or model decision) |
|
||||
|
||||
All three support both **workspace-level** (project-specific) and **global** (user-wide) scopes.
|
||||
|
||||
---
|
||||
|
||||
## Base Directory Structure
|
||||
|
||||
### Global Configuration Root
|
||||
|
||||
| OS | Path |
|
||||
|----|------|
|
||||
| **Windows** | `C:\Users\{USERNAME}\.codeium\windsurf\` |
|
||||
| **macOS** | `~/.codeium/windsurf/` |
|
||||
| **Linux** | `~/.codeium/windsurf/` |
|
||||
|
||||
### Directory Layout
|
||||
|
||||
```
|
||||
~/.codeium/windsurf/
|
||||
├── skills/ # Global skills (directories)
|
||||
│ └── {skill-name}/
|
||||
│ └── SKILL.md
|
||||
├── global_workflows/ # Global workflows (flat .md files)
|
||||
│ └── {workflow-name}.md
|
||||
├── rules/ # Global rules (flat .md files)
|
||||
│ └── {rule-name}.md
|
||||
├── memories/
|
||||
│ ├── global_rules.md # Always-on global rules (plain text)
|
||||
│ └── *.pb # Auto-generated memories (protobuf)
|
||||
├── mcp_config.json # MCP server configuration
|
||||
└── user_settings.pb # User settings (protobuf)
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## Skills
|
||||
|
||||
Skills bundle instructions with supporting resources for complex, multi-step tasks. Cascade uses **progressive disclosure** to automatically invoke skills when relevant.
|
||||
|
||||
### Storage Locations
|
||||
|
||||
| Scope | Location |
|
||||
|-------|----------|
|
||||
| **Global** | `~/.codeium/windsurf/skills/{skill-name}/SKILL.md` |
|
||||
| **Workspace** | `.windsurf/skills/{skill-name}/SKILL.md` |
|
||||
|
||||
### Directory Structure
|
||||
|
||||
Each skill is a **directory** (not a single file) containing:
|
||||
|
||||
```
|
||||
{skill-name}/
|
||||
├── SKILL.md # Required: Main skill definition
|
||||
├── references/ # Optional: Reference documentation
|
||||
├── assets/ # Optional: Images, diagrams, etc.
|
||||
├── scripts/ # Optional: Helper scripts
|
||||
└── {any-other-files} # Optional: Templates, configs, etc.
|
||||
```
|
||||
|
||||
### SKILL.md Format
|
||||
|
||||
```markdown
|
||||
---
|
||||
name: skill-name
|
||||
description: Brief description shown to model to help it decide when to invoke the skill
|
||||
---
|
||||
|
||||
# Skill Title
|
||||
|
||||
Instructions for the skill go here in markdown format.
|
||||
|
||||
## Section 1
|
||||
Step-by-step guidance...
|
||||
|
||||
## Section 2
|
||||
Reference supporting files using relative paths:
|
||||
- See [deployment-checklist.md](./deployment-checklist.md)
|
||||
- Run script: [deploy.sh](./scripts/deploy.sh)
|
||||
```
|
||||
|
||||
### Required YAML Frontmatter Fields
|
||||
|
||||
| Field | Required | Description |
|
||||
|-------|----------|-------------|
|
||||
| `name` | **Yes** | Unique identifier (lowercase letters, numbers, hyphens only). Must match directory name. |
|
||||
| `description` | **Yes** | Explains what the skill does and when to use it. Critical for automatic invocation. |
|
||||
|
||||
### Naming Convention
|
||||
|
||||
- Use **lowercase-kebab-case**: `deploy-to-staging`, `code-review`, `setup-dev-environment`
|
||||
- Name must match the directory name exactly
|
||||
|
||||
### Invocation Methods
|
||||
|
||||
1. **Automatic**: Cascade automatically invokes when request matches skill description
|
||||
2. **Manual**: Type `@skill-name` in Cascade input
|
||||
|
||||
### Example: Complete Skill
|
||||
|
||||
```
|
||||
~/.codeium/windsurf/skills/deploy-to-production/
|
||||
├── SKILL.md
|
||||
├── deployment-checklist.md
|
||||
├── rollback-procedure.md
|
||||
└── config-template.yaml
|
||||
```
|
||||
|
||||
**SKILL.md:**
|
||||
```markdown
|
||||
---
|
||||
name: deploy-to-production
|
||||
description: Guides the deployment process to production with safety checks. Use when deploying to prod, releasing, or pushing to production environment.
|
||||
---
|
||||
|
||||
## Pre-deployment Checklist
|
||||
1. Run all tests
|
||||
2. Check for uncommitted changes
|
||||
3. Verify environment variables
|
||||
|
||||
## Deployment Steps
|
||||
Follow these steps to deploy safely...
|
||||
|
||||
See [deployment-checklist.md](./deployment-checklist.md) for full checklist.
|
||||
See [rollback-procedure.md](./rollback-procedure.md) if issues occur.
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## Workflows
|
||||
|
||||
Workflows define step-by-step procedures invoked via slash commands. They guide Cascade through repetitive tasks.
|
||||
|
||||
### Storage Locations
|
||||
|
||||
| Scope | Location |
|
||||
|-------|----------|
|
||||
| **Global** | `~/.codeium/windsurf/global_workflows/{workflow-name}.md` |
|
||||
| **Workspace** | `.windsurf/workflows/{workflow-name}.md` |
|
||||
|
||||
### File Format
|
||||
|
||||
Workflows are **single markdown files** (not directories):
|
||||
|
||||
```markdown
|
||||
---
|
||||
description: Short description of what the workflow does
|
||||
---
|
||||
|
||||
# Workflow Title
|
||||
|
||||
> Arguments: [optional arguments description]
|
||||
|
||||
Step-by-step instructions in markdown.
|
||||
|
||||
1. First step
|
||||
2. Second step
|
||||
3. Third step
|
||||
```
|
||||
|
||||
### Required YAML Frontmatter Fields
|
||||
|
||||
| Field | Required | Description |
|
||||
|-------|----------|-------------|
|
||||
| `description` | **Yes** | Short title/description shown in UI |
|
||||
|
||||
### Invocation
|
||||
|
||||
- Slash command: `/workflow-name`
|
||||
- Filename becomes the command (e.g., `deploy.md` → `/deploy`)
|
||||
|
||||
### Constraints
|
||||
|
||||
- **Character limit**: 12,000 characters per workflow file
|
||||
- Workflows can call other workflows: Include instructions like "Call `/other-workflow`"
|
||||
|
||||
### Example: Complete Workflow
|
||||
|
||||
**File**: `~/.codeium/windsurf/global_workflows/address-pr-comments.md`
|
||||
|
||||
```markdown
|
||||
---
|
||||
description: Address all PR review comments systematically
|
||||
---
|
||||
|
||||
# Address PR Comments
|
||||
|
||||
> Arguments: [PR number]
|
||||
|
||||
1. Check out the PR branch: `gh pr checkout [id]`
|
||||
|
||||
2. Get comments on PR:
|
||||
```bash
|
||||
gh api --paginate repos/[owner]/[repo]/pulls/[id]/comments | jq '.[] | {user: .user.login, body, path, line}'
|
||||
```
|
||||
|
||||
3. For EACH comment:
|
||||
a. Print: "(index). From [user] on [file]:[lines] — [body]"
|
||||
b. Analyze the file and line range
|
||||
c. If unclear, ask for clarification
|
||||
d. Make the change before moving to next comment
|
||||
|
||||
4. Summarize what was done and which comments need attention
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## Rules
|
||||
|
||||
Rules provide persistent behavioral guidelines that influence how Cascade responds.
|
||||
|
||||
### Storage Locations
|
||||
|
||||
| Scope | Location |
|
||||
|-------|----------|
|
||||
| **Global** | `~/.codeium/windsurf/rules/{rule-name}.md` |
|
||||
| **Workspace** | `.windsurf/rules/{rule-name}.md` |
|
||||
|
||||
### File Format
|
||||
|
||||
Rules are **single markdown files**:
|
||||
|
||||
```markdown
|
||||
---
|
||||
description: When to use this rule
|
||||
trigger: activation_mode
|
||||
globs: ["*.py", "src/**/*.ts"]
|
||||
---
|
||||
|
||||
Rule instructions in markdown format.
|
||||
|
||||
- Guideline 1
|
||||
- Guideline 2
|
||||
- Guideline 3
|
||||
```
|
||||
|
||||
### YAML Frontmatter Fields
|
||||
|
||||
| Field | Required | Description |
|
||||
|-------|----------|-------------|
|
||||
| `description` | **Yes** | Describes when to use the rule |
|
||||
| `trigger` | Optional | Activation mode (see below) |
|
||||
| `globs` | Optional | File patterns for glob trigger |
|
||||
|
||||
### Activation Modes (trigger field)
|
||||
|
||||
| Mode | Value | Description |
|
||||
|------|-------|-------------|
|
||||
| **Manual** | `manual` | Activated via `@mention` in Cascade input |
|
||||
| **Always On** | `always` | Always applied to every conversation |
|
||||
| **Model Decision** | `model_decision` | Model decides based on description |
|
||||
| **Glob** | `glob` | Applied when working with files matching pattern |
|
||||
|
||||
### Constraints
|
||||
|
||||
- **Character limit**: 12,000 characters per rule file
|
||||
|
||||
### Example: Complete Rule
|
||||
|
||||
**File**: `~/.codeium/windsurf/rules/python-style.md`
|
||||
|
||||
```markdown
|
||||
---
|
||||
description: Python coding standards and style guidelines. Use when writing or reviewing Python code.
|
||||
trigger: glob
|
||||
globs: ["*.py", "**/*.py"]
|
||||
---
|
||||
|
||||
# Python Coding Guidelines
|
||||
|
||||
- Use type hints for all function parameters and return values
|
||||
- Follow PEP 8 style guide
|
||||
- Use early returns when possible
|
||||
- Always add docstrings to public functions and classes
|
||||
- Prefer f-strings over .format() or % formatting
|
||||
- Use pathlib instead of os.path for file operations
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## Memories
|
||||
|
||||
### Global Rules (Always-On)
|
||||
|
||||
**Location**: `~/.codeium/windsurf/memories/global_rules.md`
|
||||
|
||||
This is a special file for rules that **always apply** to all conversations. Unlike rules in the `rules/` directory, this file:
|
||||
|
||||
- Does **not** require YAML frontmatter
|
||||
- Is plain text/markdown
|
||||
- Is always active (no trigger configuration)
|
||||
|
||||
**Format:**
|
||||
```markdown
|
||||
Plain text rules that always apply to all conversations.
|
||||
|
||||
- Rule 1
|
||||
- Rule 2
|
||||
- Rule 3
|
||||
```
|
||||
|
||||
### Auto-Generated Memories
|
||||
|
||||
Cascade automatically creates memories during conversations, stored as `.pb` (protobuf) files in `~/.codeium/windsurf/memories/`. These are managed by Windsurf and should not be manually edited.
|
||||
|
||||
---
|
||||
|
||||
## System-Level Configuration (Enterprise)
|
||||
|
||||
Enterprise organizations can deploy system-level configurations that apply globally and cannot be modified by end users.
|
||||
|
||||
### System-Level Paths
|
||||
|
||||
| Type | Windows | macOS | Linux/WSL |
|
||||
|------|---------|-------|-----------|
|
||||
| **Rules** | `C:\ProgramData\Windsurf\rules\*.md` | `/Library/Application Support/Windsurf/rules/*.md` | `/etc/windsurf/rules/*.md` |
|
||||
| **Workflows** | `C:\ProgramData\Windsurf\workflows\*.md` | `/Library/Application Support/Windsurf/workflows/*.md` | `/etc/windsurf/workflows/*.md` |
|
||||
|
||||
### Precedence Order
|
||||
|
||||
When items with the same name exist at multiple levels:
|
||||
|
||||
1. **System** (highest priority) - Organization-wide, deployed by IT
|
||||
2. **Workspace** - Project-specific in `.windsurf/`
|
||||
3. **Global** - User-defined in `~/.codeium/windsurf/`
|
||||
4. **Built-in** - Default items provided by Windsurf
|
||||
|
||||
---
|
||||
|
||||
## Programmatic Creation Reference
|
||||
|
||||
### Quick Reference Table
|
||||
|
||||
| Type | Path Pattern | Format | Key Fields |
|
||||
|------|--------------|--------|------------|
|
||||
| **Skill** | `skills/{name}/SKILL.md` | YAML frontmatter + markdown | `name`, `description` |
|
||||
| **Workflow** | `global_workflows/{name}.md` (global) or `workflows/{name}.md` (workspace) | YAML frontmatter + markdown | `description` |
|
||||
| **Rule** | `rules/{name}.md` | YAML frontmatter + markdown | `description`, `trigger`, `globs` |
|
||||
| **Global Rules** | `memories/global_rules.md` | Plain text/markdown | None |
|
||||
|
||||
### Minimal Templates
|
||||
|
||||
#### Skill (SKILL.md)
|
||||
```markdown
|
||||
---
|
||||
name: my-skill
|
||||
description: What this skill does and when to use it
|
||||
---
|
||||
|
||||
Instructions here.
|
||||
```
|
||||
|
||||
#### Workflow
|
||||
```markdown
|
||||
---
|
||||
description: What this workflow does
|
||||
---
|
||||
|
||||
1. Step one
|
||||
2. Step two
|
||||
```
|
||||
|
||||
#### Rule
|
||||
```markdown
|
||||
---
|
||||
description: When this rule applies
|
||||
trigger: model_decision
|
||||
---
|
||||
|
||||
- Guideline one
|
||||
- Guideline two
|
||||
```
|
||||
|
||||
### Validation Checklist
|
||||
|
||||
When programmatically creating items:
|
||||
|
||||
- [ ] **Skills**: Directory exists with `SKILL.md` inside
|
||||
- [ ] **Skills**: `name` field matches directory name exactly
|
||||
- [ ] **Skills**: Name uses only lowercase letters, numbers, hyphens
|
||||
- [ ] **Workflows/Rules**: File is `.md` extension
|
||||
- [ ] **All**: YAML frontmatter uses `---` delimiters
|
||||
- [ ] **All**: `description` field is present and meaningful
|
||||
- [ ] **All**: File size under 12,000 characters (workflows/rules)
|
||||
|
||||
---
|
||||
|
||||
## Best Practices
|
||||
|
||||
### Writing Effective Descriptions
|
||||
|
||||
The `description` field is critical for automatic invocation. Be specific:
|
||||
|
||||
**Good:**
|
||||
```yaml
|
||||
description: Guides deployment to staging environment with pre-flight checks. Use when deploying to staging, testing releases, or preparing for production.
|
||||
```
|
||||
|
||||
**Bad:**
|
||||
```yaml
|
||||
description: Deployment stuff
|
||||
```
|
||||
|
||||
### Formatting Guidelines
|
||||
|
||||
- Use bullet points and numbered lists (easier for Cascade to follow)
|
||||
- Use markdown headers to organize sections
|
||||
- Keep rules concise and specific
|
||||
- Avoid generic rules like "write good code" (already built-in)
|
||||
|
||||
### XML Tags for Grouping
|
||||
|
||||
XML tags can effectively group related rules:
|
||||
|
||||
```markdown
|
||||
<coding_guidelines>
|
||||
- Use early returns when possible
|
||||
- Always add documentation for new functions
|
||||
- Prefer composition over inheritance
|
||||
</coding_guidelines>
|
||||
|
||||
<testing_requirements>
|
||||
- Write unit tests for all public methods
|
||||
- Maintain 80% code coverage
|
||||
</testing_requirements>
|
||||
```
|
||||
|
||||
### Skills vs Rules vs Workflows
|
||||
|
||||
| Use Case | Recommended |
|
||||
|----------|-------------|
|
||||
| Multi-step procedure with supporting files | **Skill** |
|
||||
| Repeatable CLI/automation sequence | **Workflow** |
|
||||
| Coding style preferences | **Rule** |
|
||||
| Project conventions | **Rule** |
|
||||
| Deployment procedure | **Skill** or **Workflow** |
|
||||
| Code review checklist | **Skill** |
|
||||
|
||||
---
|
||||
|
||||
## Additional Resources
|
||||
|
||||
- **Official Documentation**: [docs.windsurf.com](https://docs.windsurf.com)
|
||||
- **Skills Specification**: [agentskills.io](https://agentskills.io/home)
|
||||
- **Rule Templates**: [windsurf.com/editor/directory](https://windsurf.com/editor/directory)
|
||||
16
package.json
16
package.json
@@ -1,11 +1,13 @@
|
||||
{
|
||||
"name": "@every-env/compound-plugin",
|
||||
"version": "0.7.0",
|
||||
"version": "2.42.0",
|
||||
"type": "module",
|
||||
"private": false,
|
||||
"bin": {
|
||||
"compound-plugin": "./src/index.ts"
|
||||
"compound-plugin": "src/index.ts"
|
||||
},
|
||||
"homepage": "https://github.com/EveryInc/compound-engineering-plugin",
|
||||
"repository": "https://github.com/EveryInc/compound-engineering-plugin",
|
||||
"publishConfig": {
|
||||
"access": "public"
|
||||
},
|
||||
@@ -14,13 +16,19 @@
|
||||
"convert": "bun run src/index.ts convert",
|
||||
"list": "bun run src/index.ts list",
|
||||
"cli:install": "bun run src/index.ts install",
|
||||
"test": "bun test"
|
||||
"test": "bun test",
|
||||
"release:preview": "bun run scripts/release/preview.ts",
|
||||
"release:sync-metadata": "bun run scripts/release/sync-metadata.ts --write",
|
||||
"release:validate": "bun run scripts/release/validate.ts"
|
||||
},
|
||||
"dependencies": {
|
||||
"citty": "^0.1.6",
|
||||
"js-yaml": "^4.1.0"
|
||||
},
|
||||
"devDependencies": {
|
||||
"bun-types": "^1.0.0"
|
||||
"@semantic-release/changelog": "^6.0.3",
|
||||
"@semantic-release/git": "^10.0.1",
|
||||
"bun-types": "^1.0.0",
|
||||
"semantic-release": "^25.0.3"
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,102 +0,0 @@
|
||||
# Grow Your Own Garden: Adaptive Agent Ecosystem
|
||||
|
||||
> **Issue:** https://github.com/EveryInc/compound-engineering-plugin/issues/20
|
||||
|
||||
## The Idea
|
||||
|
||||
Everyone grows their own garden, but we're all using the same process.
|
||||
|
||||
Start from a **seed** (minimal core: `/plan`, `/work`, `/review`, `/compound`). Each `/compound` loop can suggest adding agents based on what you're working on—like building up a test suite to prevent regressions, but for code review expertise.
|
||||
|
||||
## Current Problem
|
||||
|
||||
- Monolithic plugin: 24 agents, users use ~30%
|
||||
- No personalization (same agents for Rails dev and Python dev)
|
||||
- Static collection that doesn't adapt
|
||||
|
||||
## Proposed Solution
|
||||
|
||||
### The Seed (Core Plugin)
|
||||
|
||||
4 commands + minimal agents:
|
||||
|
||||
| Component | What's Included |
|
||||
|-----------|-----------------|
|
||||
| Commands | `/plan`, `/work`, `/review`, `/compound` |
|
||||
| Review Agents | security, performance, simplicity, architecture, patterns |
|
||||
| Research Agents | best-practices, framework-docs, git-history, repo-analyst |
|
||||
| Skills | compound-docs, file-todos, git-worktree |
|
||||
| MCP Servers | playwright, context7 |
|
||||
|
||||
### The Growth Loop
|
||||
|
||||
After each `/compound`:
|
||||
|
||||
```
|
||||
✅ Learning documented
|
||||
|
||||
💡 It looks like you're using Rails.
|
||||
Would you like to add the "DHH Rails Reviewer"?
|
||||
|
||||
[y] Yes [n] No [x] Never ask
|
||||
```
|
||||
|
||||
Three sources of new agents:
|
||||
1. **Predefined** - "You're using Rails, add DHH reviewer?"
|
||||
2. **Dynamic** - "You're using actor model, create an expert?"
|
||||
3. **Custom** - "Want to create an agent for this pattern?"
|
||||
|
||||
### Agent Storage
|
||||
|
||||
```
|
||||
.claude/agents/ → Project-specific (highest priority)
|
||||
~/.claude/agents/ → User's garden
|
||||
plugin/agents/ → From installed plugins
|
||||
```
|
||||
|
||||
## Implementation Phases
|
||||
|
||||
### Phase 1: Split the Plugin
|
||||
- Create `agent-library/` with framework-specific agents (Rails, Python, TypeScript, Frontend)
|
||||
- Keep `compound-engineering` as core with universal agents
|
||||
- No breaking changes—existing users unaffected
|
||||
|
||||
### Phase 2: Agent Discovery
|
||||
- `/review` discovers agents from all three locations
|
||||
- Project agents override user agents override plugin agents
|
||||
|
||||
### Phase 3: Growth via /compound
|
||||
- Detect tech stack (Gemfile, package.json, etc.)
|
||||
- Suggest relevant agents after documenting learnings
|
||||
- Install accepted agents to `~/.claude/agents/`
|
||||
|
||||
### Phase 4: Management
|
||||
- `/agents list` - See your garden
|
||||
- `/agents add <name>` - Add from library
|
||||
- `/agents disable <name>` - Temporarily disable
|
||||
|
||||
## What Goes Where
|
||||
|
||||
**Core (seed):** 11 framework-agnostic agents
|
||||
- security-sentinel, performance-oracle, code-simplicity-reviewer
|
||||
- architecture-strategist, pattern-recognition-specialist
|
||||
- 4 research agents, 2 workflow agents
|
||||
|
||||
**Agent Library:** 10 specialized agents
|
||||
- Rails: kieran-rails, dhh-rails, data-integrity (3)
|
||||
- Python: kieran-python (1)
|
||||
- TypeScript: kieran-typescript (1)
|
||||
- Frontend: julik-races, design-iterator, design-reviewer, figma-sync (4)
|
||||
- Editorial: every-style-editor (1)
|
||||
|
||||
## Key Constraint
|
||||
|
||||
Claude Code doesn't support plugin dependencies. Each plugin must be independent. Users manually install what they need, or we suggest additions via `/compound`.
|
||||
|
||||
## Acceptance Criteria
|
||||
|
||||
- [ ] Core plugin works standalone with universal agents
|
||||
- [ ] `/compound` suggests agents based on detected tech stack
|
||||
- [ ] Users can accept/decline suggestions
|
||||
- [ ] `/agents` command for garden management
|
||||
- [ ] No breaking changes for existing users
|
||||
@@ -1,279 +0,0 @@
|
||||
# Landing Page LaunchKit Refresh
|
||||
|
||||
## Overview
|
||||
|
||||
Review and enhance the `/docs/index.html` landing page using LaunchKit elements and Pragmatic Technical Writing style (Hunt/Thomas, Joel Spolsky). The current implementation is strong but can be refined section-by-section.
|
||||
|
||||
## Current State Assessment
|
||||
|
||||
### What's Working Well
|
||||
- Specific, outcome-focused hero headline ("12 expert opinions in 30 seconds")
|
||||
- Developer-authentic copywriting (N+1 queries, CORS, SQL injection)
|
||||
- Stats section with clear metrics (23 agents, 16 commands, 11 skills, 2 MCP servers)
|
||||
- Philosophy section with concrete story (N+1 query bug)
|
||||
- Three-step installation with actual commands
|
||||
- FAQ accordion following LaunchKit patterns
|
||||
- Categorized feature sections with code examples
|
||||
|
||||
### Missing Elements (From Best Practices Research)
|
||||
1. **Social Proof Section** - No testimonials, GitHub stars, or user metrics
|
||||
2. **Visual Demo** - No GIF/animation showing the tool in action
|
||||
3. **Arrow icons on CTAs** - 26% conversion boost from studies
|
||||
4. **Trust indicators** - Open source badge, license info
|
||||
|
||||
---
|
||||
|
||||
## Section-by-Section Review Plan
|
||||
|
||||
### 1. Hero Section (lines 56-78)
|
||||
|
||||
**Current:**
|
||||
```html
|
||||
<h1>Your Code Reviews Just Got 12 Expert Opinions. In 30 Seconds.</h1>
|
||||
```
|
||||
|
||||
**Review Checklist:**
|
||||
- [ ] Headline follows Pragmatic Writing (concrete before abstract) ✅
|
||||
- [ ] Eyebrow badge is current (Version 2.6.0) - verify
|
||||
- [ ] Description paragraph under 3 sentences ✅
|
||||
- [ ] Button group has arrow icon on primary CTA
|
||||
- [ ] "Read the Docs" secondary CTA present ✅
|
||||
|
||||
**Potential Improvements:**
|
||||
- Add `→` arrow to "Install Plugin" button
|
||||
- Consider adding animated terminal GIF below buttons showing `/review` in action
|
||||
|
||||
### 2. Stats Section (lines 81-104)
|
||||
|
||||
**Current:** 4 stat cards (23 agents, 16 commands, 11 skills, 2 MCP servers)
|
||||
|
||||
**Review Checklist:**
|
||||
- [ ] Numbers are accurate (verify against actual file counts)
|
||||
- [ ] Icons are appropriate for each stat
|
||||
- [ ] Hover effects working properly
|
||||
- [ ] Mobile layout (2x2 grid) is readable
|
||||
|
||||
**Potential Improvements:**
|
||||
- Add "developers using" or "reviews run" metric if available
|
||||
- Consider adding subtle animation on scroll
|
||||
|
||||
### 3. Philosophy Section (lines 107-192)
|
||||
|
||||
**Current:** "Why Your Third Code Review Should Be Easier Than Your First" with N+1 query story
|
||||
|
||||
**Review Checklist:**
|
||||
- [ ] Opens with concrete story (N+1 query) ✅
|
||||
- [ ] Quote block is memorable and quotable
|
||||
- [ ] Four pillars (Plan, Delegate, Assess, Codify) are clear
|
||||
- [ ] Each pillar has: tagline, description, tool tags
|
||||
- [ ] Descriptions use "you" voice ✅
|
||||
|
||||
**Potential Improvements:**
|
||||
- Review pillar descriptions for passive voice
|
||||
- Ensure each pillar description follows PAS (Problem, Agitate, Solve) pattern
|
||||
- Check tool tags are accurate and current
|
||||
|
||||
### 4. Agents Section (lines 195-423)
|
||||
|
||||
**Current:** 23 agents in 5 categories (Review, Research, Design, Workflow, Docs)
|
||||
|
||||
**Review Checklist:**
|
||||
- [ ] All 23 agents are listed (count actual files)
|
||||
- [ ] Categories are logical and scannable
|
||||
- [ ] Each card has: name, badge, description, usage code
|
||||
- [ ] Descriptions are conversational (not passive)
|
||||
- [ ] Critical badges (Security, Data) stand out
|
||||
|
||||
**Potential Improvements:**
|
||||
- Review agent descriptions against pragmatic writing checklist
|
||||
- Ensure descriptions answer "when would I use this?"
|
||||
- Add concrete scenarios to generic descriptions
|
||||
|
||||
### 5. Commands Section (lines 426-561)
|
||||
|
||||
**Current:** 16 commands in 2 categories (Workflow, Utility)
|
||||
|
||||
**Review Checklist:**
|
||||
- [ ] All 16 commands are listed (count actual files)
|
||||
- [ ] Core workflow commands are highlighted
|
||||
- [ ] Descriptions are action-oriented
|
||||
- [ ] Command names match actual implementation
|
||||
|
||||
**Potential Improvements:**
|
||||
- Review command descriptions for passive voice
|
||||
- Lead with outcomes, not features
|
||||
- Add "saves you X minutes" framing where appropriate
|
||||
|
||||
### 6. Skills Section (lines 564-703)
|
||||
|
||||
**Current:** 11 skills in 3 categories (Development, Content/Workflow, Image Generation)
|
||||
|
||||
**Review Checklist:**
|
||||
- [ ] All 11 skills are listed (count actual directories)
|
||||
- [ ] Featured skill (gemini-imagegen) is properly highlighted
|
||||
- [ ] API key requirement is clear
|
||||
- [ ] Skill invocation syntax is correct
|
||||
|
||||
**Potential Improvements:**
|
||||
- Review skill descriptions against pragmatic writing
|
||||
- Ensure each skill answers "what problem does this solve?"
|
||||
|
||||
### 7. MCP Servers Section (lines 706-751)
|
||||
|
||||
**Current:** 2 MCP servers (Playwright, Context7)
|
||||
|
||||
**Review Checklist:**
|
||||
- [ ] Tool lists are accurate
|
||||
- [ ] Descriptions explain WHY not just WHAT
|
||||
- [ ] Framework support list is current (100+)
|
||||
|
||||
**Potential Improvements:**
|
||||
- Add concrete example of each server in action
|
||||
- Consider before/after comparison
|
||||
|
||||
### 8. Installation Section (lines 754-798)
|
||||
|
||||
**Current:** "Three Commands. Zero Configuration." with 3 steps
|
||||
|
||||
**Review Checklist:**
|
||||
- [ ] Commands are accurate and work
|
||||
- [ ] Step 3 shows actual usage examples
|
||||
- [ ] Timeline visual (vertical line) renders correctly
|
||||
- [ ] Copy buttons work on code blocks
|
||||
|
||||
**Potential Improvements:**
|
||||
- Add copy-to-clipboard functionality if missing
|
||||
- Consider adding "What you'll see" output example
|
||||
|
||||
### 9. FAQ Section (lines 801-864)
|
||||
|
||||
**Current:** 5 questions in accordion format
|
||||
|
||||
**Review Checklist:**
|
||||
- [ ] Questions address real objections
|
||||
- [ ] Answers are conversational (use "you")
|
||||
- [ ] Accordion expand/collapse works
|
||||
- [ ] No passive voice in answers
|
||||
|
||||
**Potential Improvements:**
|
||||
- Review for weasel words ("best practices suggest")
|
||||
- Ensure answers are direct and actionable
|
||||
|
||||
### 10. CTA Section (lines 868-886)
|
||||
|
||||
**Current:** "Install Once. Compound Forever." with Install + GitHub buttons
|
||||
|
||||
**Review Checklist:**
|
||||
- [ ] Badge is eye-catching ("Free & Open Source")
|
||||
- [ ] Headline restates core value proposition
|
||||
- [ ] Primary CTA has arrow icon ✅
|
||||
- [ ] Trust line at bottom
|
||||
|
||||
**Potential Improvements:**
|
||||
- Review trust line copy
|
||||
- Consider adding social proof element
|
||||
|
||||
---
|
||||
|
||||
## NEW: Social Proof Section (To Add)
|
||||
|
||||
**Position:** After Stats section, before Philosophy section
|
||||
|
||||
**Components:**
|
||||
- GitHub stars counter (dynamic or static)
|
||||
- "Trusted by X developers" metric
|
||||
- 2-3 testimonial quotes (if available)
|
||||
- Company logos (if applicable)
|
||||
|
||||
**LaunchKit Pattern:**
|
||||
```html
|
||||
<section class="social-proof-section">
|
||||
<div class="heading centered">
|
||||
<p class="paragraph m secondary">Trusted by developers at</p>
|
||||
</div>
|
||||
<div class="logo-grid">
|
||||
<!-- Company logos or GitHub badge -->
|
||||
</div>
|
||||
</section>
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## Pragmatic Writing Style Checklist (Apply to ALL Copy)
|
||||
|
||||
### The Five Laws
|
||||
1. **Concrete Before Abstract** - Story/example first, then principle
|
||||
2. **Physical Analogies** - Import metaphors readers understand
|
||||
3. **Conversational Register** - Use "you", contractions, asides
|
||||
4. **Numbered Frameworks** - Create referenceable structures
|
||||
5. **Humor as Architecture** - Mental anchors for dense content
|
||||
|
||||
### Anti-Patterns to Find and Fix
|
||||
- [ ] "It is recommended that..." → "Do this:"
|
||||
- [ ] "Best practices suggest..." → "Here's what works:"
|
||||
- [ ] Passive voice → Active voice
|
||||
- [ ] Abstract claims → Specific examples
|
||||
- [ ] Walls of text → Scannable lists
|
||||
|
||||
### Quality Checklist (Per Section)
|
||||
- [ ] Opens with concrete story or example?
|
||||
- [ ] Can reader skim headers and get the arc?
|
||||
- [ ] Uses "you" at least once?
|
||||
- [ ] Clear action reader can take?
|
||||
- [ ] Reads aloud like speech?
|
||||
|
||||
---
|
||||
|
||||
## Implementation Phases
|
||||
|
||||
### Phase 1: Copy Audit (No HTML Changes)
|
||||
1. Read through entire page
|
||||
2. Flag passive voice instances
|
||||
3. Flag abstract claims without examples
|
||||
4. Flag missing "you" voice
|
||||
5. Document improvements needed
|
||||
|
||||
### Phase 2: Copy Rewrites
|
||||
1. Rewrite flagged sections following pragmatic style
|
||||
2. Ensure each section passes quality checklist
|
||||
3. Maintain existing HTML structure
|
||||
|
||||
### Phase 3: Component Additions
|
||||
1. Add arrow icons to primary CTAs
|
||||
2. Add social proof section (if data available)
|
||||
3. Consider visual demo element
|
||||
|
||||
### Phase 4: Verification
|
||||
1. Validate all counts (agents, commands, skills)
|
||||
2. Test all links and buttons
|
||||
3. Verify mobile responsiveness
|
||||
4. Check accessibility
|
||||
|
||||
---
|
||||
|
||||
## Files to Modify
|
||||
|
||||
| File | Changes |
|
||||
|------|---------|
|
||||
| `docs/index.html` | Copy rewrites, potential new section |
|
||||
| `docs/css/style.css` | Social proof styles (if adding) |
|
||||
|
||||
---
|
||||
|
||||
## Success Criteria
|
||||
|
||||
1. All copy passes Pragmatic Writing quality checklist
|
||||
2. No passive voice in any description
|
||||
3. Every feature section answers "why should I care?"
|
||||
4. Stats are accurate against actual file counts
|
||||
5. Page loads in <3 seconds
|
||||
6. Mobile layout is fully functional
|
||||
|
||||
---
|
||||
|
||||
## References
|
||||
|
||||
- LaunchKit Template: https://launchkit.evilmartians.io/
|
||||
- Pragmatic Writing Skill: `~/.claude/skills/pragmatic-writing-skill/SKILL.md`
|
||||
- Current Landing Page: `/Users/kieranklaassen/every-marketplace/docs/index.html`
|
||||
- Style CSS: `/Users/kieranklaassen/every-marketplace/docs/css/style.css`
|
||||
21
plugins/coding-tutor/.cursor-plugin/plugin.json
Normal file
21
plugins/coding-tutor/.cursor-plugin/plugin.json
Normal file
@@ -0,0 +1,21 @@
|
||||
{
|
||||
"name": "coding-tutor",
|
||||
"displayName": "Coding Tutor",
|
||||
"version": "1.2.1",
|
||||
"description": "Personalized coding tutorials that use your actual codebase for examples with spaced repetition quizzes",
|
||||
"author": {
|
||||
"name": "Nityesh Agarwal"
|
||||
},
|
||||
"homepage": "https://github.com/EveryInc/compound-engineering-plugin",
|
||||
"repository": "https://github.com/EveryInc/compound-engineering-plugin",
|
||||
"license": "MIT",
|
||||
"keywords": [
|
||||
"cursor",
|
||||
"plugin",
|
||||
"coding",
|
||||
"programming",
|
||||
"tutorial",
|
||||
"learning",
|
||||
"spaced-repetition"
|
||||
]
|
||||
}
|
||||
@@ -1,7 +1,7 @@
|
||||
{
|
||||
"name": "compound-engineering",
|
||||
"version": "2.34.0",
|
||||
"description": "AI-powered development tools. 29 agents, 22 commands, 19 skills, 1 MCP server for code review, research, design, and workflow automation.",
|
||||
"version": "2.42.0",
|
||||
"description": "AI-powered development tools. 29 agents, 44 skills, 1 MCP server for code review, research, design, and workflow automation.",
|
||||
"author": {
|
||||
"name": "Kieran Klaassen",
|
||||
"email": "kieran@every.to",
|
||||
|
||||
31
plugins/compound-engineering/.cursor-plugin/plugin.json
Normal file
31
plugins/compound-engineering/.cursor-plugin/plugin.json
Normal file
@@ -0,0 +1,31 @@
|
||||
{
|
||||
"name": "compound-engineering",
|
||||
"displayName": "Compound Engineering",
|
||||
"version": "2.42.0",
|
||||
"description": "AI-powered development tools. 29 agents, 44 skills, 1 MCP server for code review, research, design, and workflow automation.",
|
||||
"author": {
|
||||
"name": "Kieran Klaassen",
|
||||
"email": "kieran@every.to",
|
||||
"url": "https://github.com/kieranklaassen"
|
||||
},
|
||||
"homepage": "https://every.to/source-code/my-ai-had-already-fixed-the-code-before-i-saw-it",
|
||||
"repository": "https://github.com/EveryInc/compound-engineering-plugin",
|
||||
"license": "MIT",
|
||||
"keywords": [
|
||||
"cursor",
|
||||
"plugin",
|
||||
"ai-powered",
|
||||
"compound-engineering",
|
||||
"workflow-automation",
|
||||
"code-review",
|
||||
"rails",
|
||||
"ruby",
|
||||
"python",
|
||||
"typescript",
|
||||
"knowledge-management",
|
||||
"image-generation",
|
||||
"agent-browser",
|
||||
"browser-automation"
|
||||
],
|
||||
"mcpServers": ".mcp.json"
|
||||
}
|
||||
11
plugins/compound-engineering/.mcp.json
Normal file
11
plugins/compound-engineering/.mcp.json
Normal file
@@ -0,0 +1,11 @@
|
||||
{
|
||||
"mcpServers": {
|
||||
"context7": {
|
||||
"type": "http",
|
||||
"url": "https://mcp.context7.com/mcp",
|
||||
"headers": {
|
||||
"x-api-key": "${CONTEXT7_API_KEY:-}"
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
130
plugins/compound-engineering/AGENTS.md
Normal file
130
plugins/compound-engineering/AGENTS.md
Normal file
@@ -0,0 +1,130 @@
|
||||
# Plugin Instructions
|
||||
|
||||
These instructions apply when working under `plugins/compound-engineering/`.
|
||||
They supplement the repo-root `AGENTS.md`.
|
||||
|
||||
# Compounding Engineering Plugin Development
|
||||
|
||||
## Versioning Requirements
|
||||
|
||||
**IMPORTANT**: Routine PRs should not cut releases for this plugin.
|
||||
|
||||
The repo uses an automated release process to prepare plugin releases, including version selection and changelog generation. Because multiple PRs may merge before the next release, contributors cannot know the final released version from within an individual PR.
|
||||
|
||||
### Contributor Rules
|
||||
|
||||
- Do **not** manually bump `.claude-plugin/plugin.json` version in a normal feature PR.
|
||||
- Do **not** manually bump `.claude-plugin/marketplace.json` plugin version in a normal feature PR.
|
||||
- Do **not** cut a release section in the canonical root `CHANGELOG.md` for a normal feature PR.
|
||||
- Do update substantive docs that are part of the actual change, such as `README.md`, component tables, usage instructions, or counts when they would otherwise become inaccurate.
|
||||
|
||||
### Pre-Commit Checklist
|
||||
|
||||
Before committing ANY changes:
|
||||
|
||||
- [ ] No manual release-version bump in `.claude-plugin/plugin.json`
|
||||
- [ ] No manual release-version bump in `.claude-plugin/marketplace.json`
|
||||
- [ ] No manual release entry added to the root `CHANGELOG.md`
|
||||
- [ ] README.md component counts verified
|
||||
- [ ] README.md tables accurate (agents, commands, skills)
|
||||
- [ ] plugin.json description matches current counts
|
||||
|
||||
### Directory Structure
|
||||
|
||||
```
|
||||
agents/
|
||||
├── review/ # Code review agents
|
||||
├── research/ # Research and analysis agents
|
||||
├── design/ # Design and UI agents
|
||||
└── docs/ # Documentation agents
|
||||
|
||||
skills/
|
||||
├── ce-*/ # Core workflow skills (ce:plan, ce:review, etc.)
|
||||
└── */ # All other skills
|
||||
```
|
||||
|
||||
> **Note:** Commands were migrated to skills in v2.39.0. All former
|
||||
> `/command-name` slash commands now live under `skills/command-name/SKILL.md`
|
||||
> and work identically in Claude Code. Other targets may convert or map these references differently.
|
||||
|
||||
## Command Naming Convention
|
||||
|
||||
**Workflow commands** use `ce:` prefix to unambiguously identify them as compound-engineering commands:
|
||||
- `/ce:brainstorm` - Explore requirements and approaches before planning
|
||||
- `/ce:plan` - Create implementation plans
|
||||
- `/ce:review` - Run comprehensive code reviews
|
||||
- `/ce:work` - Execute work items systematically
|
||||
- `/ce:compound` - Document solved problems
|
||||
|
||||
**Why `ce:`?** Claude Code has built-in `/plan` and `/review` commands. The `ce:` namespace (short for compound-engineering) makes it immediately clear these commands belong to this plugin.
|
||||
|
||||
## Skill Compliance Checklist
|
||||
|
||||
When adding or modifying skills, verify compliance with the skill spec:
|
||||
|
||||
### YAML Frontmatter (Required)
|
||||
|
||||
- [ ] `name:` present and matches directory name (lowercase-with-hyphens)
|
||||
- [ ] `description:` present and describes **what it does and when to use it** (per official spec: "Explains code with diagrams. Use when exploring how code works.")
|
||||
|
||||
### Reference Links (Required if references/ exists)
|
||||
|
||||
- [ ] All files in `references/` are linked as `[filename.md](./references/filename.md)`
|
||||
- [ ] All files in `assets/` are linked as `[filename](./assets/filename)`
|
||||
- [ ] All files in `scripts/` are linked as `[filename](./scripts/filename)`
|
||||
- [ ] No bare backtick references like `` `references/file.md` `` - use proper markdown links
|
||||
|
||||
### Writing Style
|
||||
|
||||
- [ ] Use imperative/infinitive form (verb-first instructions)
|
||||
- [ ] Avoid second person ("you should") - use objective language ("To accomplish X, do Y")
|
||||
|
||||
### Cross-Platform User Interaction
|
||||
|
||||
- [ ] When a skill needs to ask the user a question, instruct use of the platform's blocking question tool and name the known equivalents (`AskUserQuestion` in Claude Code, `request_user_input` in Codex, `ask_user` in Gemini)
|
||||
- [ ] Include a fallback for environments without a question tool (e.g., present numbered options and wait for the user's reply before proceeding)
|
||||
|
||||
### Cross-Platform Reference Rules
|
||||
|
||||
This plugin is authored once, then converted for other agent platforms. Commands and agents are transformed during that conversion, but `plugin.skills` are usually copied almost exactly as written.
|
||||
|
||||
- [ ] Because of that, slash references inside command or agent content are acceptable when they point to real published commands; target-specific conversion can remap them.
|
||||
- [ ] Inside a pass-through `SKILL.md`, do not assume slash references will be remapped for another platform. Write references according to what will still make sense after the skill is copied as-is.
|
||||
- [ ] When one skill refers to another skill, prefer semantic wording such as "load the `document-review` skill" rather than slash syntax.
|
||||
- [ ] Use slash syntax only when referring to an actual published command or workflow such as `/ce:work` or `/deepen-plan`.
|
||||
|
||||
### Tool Selection in Agents and Skills
|
||||
|
||||
Agents and skills that explore codebases must prefer native tools over shell commands.
|
||||
|
||||
Why: shell-heavy exploration causes avoidable permission prompts in sub-agent workflows; native file-search, content-search, and file-read tools avoid that.
|
||||
|
||||
- [ ] Never instruct agents to use `find`, `ls`, `cat`, `head`, `tail`, `grep`, `rg`, `wc`, or `tree` through a shell for routine file discovery, content search, or file reading
|
||||
- [ ] Describe tools by capability class with platform hints — e.g., "Use the native file-search/glob tool (e.g., Glob in Claude Code)" — not by Claude Code-specific tool names alone
|
||||
- [ ] When shell is the only option (e.g., `ast-grep`, `bundle show`, git commands), instruct one simple command at a time — no chaining (`&&`, `||`, `;`), pipes, or redirects
|
||||
- [ ] Do not encode shell recipes for routine exploration when native tools can do the job; encode intent and preferred tool classes instead
|
||||
- [ ] For shell-only workflows (e.g., `gh`, `git`, `bundle show`, project CLIs), explicit command examples are acceptable when they are simple, task-scoped, and not chained together
|
||||
|
||||
### Quick Validation Command
|
||||
|
||||
```bash
|
||||
# Check for unlinked references in a skill
|
||||
grep -E '`(references|assets|scripts)/[^`]+`' skills/*/SKILL.md
|
||||
# Should return nothing if all refs are properly linked
|
||||
|
||||
# Check description format - should describe what + when
|
||||
grep -E '^description:' skills/*/SKILL.md
|
||||
```
|
||||
|
||||
## Adding Components
|
||||
|
||||
- **New skill:** Create `skills/<name>/SKILL.md` with required YAML frontmatter (`name`, `description`). Reference files go in `skills/<name>/references/`.
|
||||
- **New agent:** Create `agents/<category>/<name>.md` with frontmatter. Categories: `review`, `research`, `design`, `docs`, `workflow`.
|
||||
|
||||
## Beta Skills
|
||||
|
||||
Beta skills use a `-beta` suffix and `disable-model-invocation: true` to prevent accidental auto-triggering. See `docs/solutions/skill-design/beta-skills-framework.md` for naming, validation, and promotion rules.
|
||||
|
||||
## Documentation
|
||||
|
||||
See `docs/solutions/plugin-versioning-requirements.md` for detailed versioning workflow.
|
||||
@@ -1,10 +1,131 @@
|
||||
# Changelog
|
||||
|
||||
This file is no longer the canonical changelog for compound-engineering releases.
|
||||
|
||||
Historical entries are preserved below, but new release history is recorded in the root [`CHANGELOG.md`](../../CHANGELOG.md).
|
||||
|
||||
All notable changes to the compound-engineering plugin will be documented in this file.
|
||||
|
||||
The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.0.0/),
|
||||
and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.html).
|
||||
|
||||
## [2.39.0] - 2026-03-10
|
||||
|
||||
### Added
|
||||
|
||||
- **ce:compound context budget precheck** — Warns when context is constrained and offers compact-safe mode to avoid compaction mid-compound ([#235](https://github.com/EveryInc/compound-engineering-plugin/pull/235))
|
||||
- **ce:plan daily sequence numbers** — Plan filenames now include a 3-digit daily sequence number (e.g., `2026-03-10-001-feat-...`) to prevent collisions ([#238](https://github.com/EveryInc/compound-engineering-plugin/pull/238))
|
||||
- **ce:review serial mode** — Pass `--serial` flag (or auto-detects when 6+ agents configured) to run review agents sequentially, preventing context limit crashes ([#237](https://github.com/EveryInc/compound-engineering-plugin/pull/237))
|
||||
- **agent-browser inspection & debugging commands** — Added JS eval, console/errors, network, storage, device emulation, element debugging, recording/tracing, tabs, and advanced mouse commands to agent-browser skill ([#236](https://github.com/EveryInc/compound-engineering-plugin/pull/236))
|
||||
- **test-browser port detection** — Auto-detects dev server port from CLAUDE.md, package.json, or .env files; supports `--port` flag ([#233](https://github.com/EveryInc/compound-engineering-plugin/pull/233))
|
||||
- **lfg phase gating** — Added explicit GATE checks between /lfg steps to enforce plan-before-work ordering ([#231](https://github.com/EveryInc/compound-engineering-plugin/pull/231))
|
||||
|
||||
### Fixed
|
||||
|
||||
- **Context7 API key auth** — MCP server config now passes `CONTEXT7_API_KEY` via `x-api-key` header to avoid anonymous rate limits ([#232](https://github.com/EveryInc/compound-engineering-plugin/pull/232))
|
||||
- **CLI: MCP server merge order** — `sync` now correctly overwrites same-named MCP servers with plugin values instead of preserving stale entries
|
||||
|
||||
### Removed
|
||||
|
||||
- **every-style-editor agent** — Removed duplicate agent; functionality already exists as `every-style-editor` skill ([#234](https://github.com/EveryInc/compound-engineering-plugin/pull/234))
|
||||
|
||||
### Contributors
|
||||
|
||||
- Matt Van Horn ([@mvanhorn](https://x.com/mvanhorn)) — PRs #231–#238
|
||||
|
||||
---
|
||||
|
||||
## [2.38.1] - 2026-03-01
|
||||
|
||||
### Fixed
|
||||
|
||||
- **Cross-platform `AskUserQuestion` fallback** — `setup` skill and `create-new-skill`/`add-workflow` workflows now include an "Interaction Method" preamble that instructs non-Claude LLMs (Codex, Gemini, Copilot, Kiro) to use numbered lists instead of `AskUserQuestion`, preventing silent auto-configuration. ([#204](https://github.com/EveryInc/compound-engineering-plugin/issues/204))
|
||||
- **Codex AGENTS.md `AskUserQuestion` mapping** — Strengthened from "ask the user in chat" to structured numbered-list guidance with multi-select support and a "never skip or auto-configure" rule.
|
||||
- **Skill compliance checklist** — Added `AskUserQuestion` lint rule to `CLAUDE.md` to prevent recurrence.
|
||||
|
||||
---
|
||||
|
||||
## [2.38.0] - 2026-03-01
|
||||
|
||||
### Changed
|
||||
- `workflows:plan`, `workflows:work`, `workflows:review`, `workflows:brainstorm`, `workflows:compound` renamed to `ce:plan`, `ce:work`, `ce:review`, `ce:brainstorm`, `ce:compound` for clarity — the `ce:` prefix unambiguously identifies these as compound-engineering commands
|
||||
|
||||
### Deprecated
|
||||
- `workflows:*` commands — all five remain functional as aliases that forward to their `ce:*` equivalents with a deprecation notice. Will be removed in a future version.
|
||||
|
||||
---
|
||||
|
||||
## [2.37.2] - 2026-03-01
|
||||
|
||||
### Added
|
||||
|
||||
- **CLI: auto-detect install targets** — `bunx @every-env/compound-plugin install compound-engineering --to all` auto-detects installed AI coding tools and installs to all of them in one command. ([#191](https://github.com/EveryInc/compound-engineering-plugin/pull/191))
|
||||
- **CLI: Gemini sync** — `sync --target gemini` symlinks personal skills to `.gemini/skills/` and merges MCP servers into `.gemini/settings.json`. ([#191](https://github.com/EveryInc/compound-engineering-plugin/pull/191))
|
||||
- **CLI: sync defaults to `--target all`** — Running `sync` with no target now syncs to all detected tools automatically. ([#191](https://github.com/EveryInc/compound-engineering-plugin/pull/191))
|
||||
|
||||
---
|
||||
|
||||
## [2.37.1] - 2026-03-01
|
||||
|
||||
### Fixed
|
||||
|
||||
- **`/workflows:review` rendering** — Fixed broken markdown output: "Next Steps" items 3 & 4 and Severity Breakdown no longer leak outside the Summary Report template, section numbering fixed (was jumping 5→7, now correct), removed orphaned fenced code block delimiters that caused the entire End-to-End Testing section to render as a code block, and fixed unclosed quoted string in section 1. ([#214](https://github.com/EveryInc/compound-engineering-plugin/pull/214)) — thanks [@XSAM](https://github.com/XSAM)!
|
||||
- **`.worktrees` gitignore** — Added `.worktrees/` to `.gitignore` to prevent worktree directories created by the `git-worktree` skill from being tracked. ([#213](https://github.com/EveryInc/compound-engineering-plugin/pull/213)) — thanks [@XSAM](https://github.com/XSAM)!
|
||||
|
||||
---
|
||||
|
||||
## [2.37.0] - 2026-03-01
|
||||
|
||||
### Added
|
||||
|
||||
- **`proof` skill** — Create, edit, comment on, and share markdown documents via Proof's web API and local bridge. Supports document creation, track-changes suggestions, comments, and bulk rewrites. No authentication required for creating shared documents.
|
||||
- **Optional Proof sharing in `/workflows:brainstorm`** — "Share to Proof" is now a menu option in Phase 4 handoff, letting you upload the brainstorm document when you want to, rather than automatically on every run.
|
||||
- **Optional Proof sharing in `/workflows:plan`** — "Share to Proof" is now a menu option in Post-Generation Options, letting you upload the plan file on demand rather than automatically.
|
||||
|
||||
---
|
||||
|
||||
## [2.36.0] - 2026-03-01
|
||||
|
||||
### Added
|
||||
|
||||
- **OpenClaw install target** — `bunx @every-env/compound-plugin install compound-engineering --to openclaw` now installs the plugin to OpenClaw's extensions directory. ([#217](https://github.com/EveryInc/compound-engineering-plugin/pull/217)) — thanks [@TrendpilotAI](https://github.com/TrendpilotAI)!
|
||||
- **Qwen Code install target** — `bunx @every-env/compound-plugin install compound-engineering --to qwen` now installs the plugin to Qwen Code's extensions directory. ([#220](https://github.com/EveryInc/compound-engineering-plugin/pull/220)) — thanks [@rlam3](https://github.com/rlam3)!
|
||||
- **Windsurf install target** — `bunx @every-env/compound-plugin install compound-engineering --to windsurf` converts plugins to Windsurf format. Agents become Windsurf skills, commands become flat workflows, and MCP servers write to `mcp_config.json`. Defaults to global scope (`~/.codeium/windsurf/`); use `--scope workspace` for project-level output. ([#202](https://github.com/EveryInc/compound-engineering-plugin/pull/202)) — thanks [@rburnham52](https://github.com/rburnham52)!
|
||||
|
||||
### Fixed
|
||||
|
||||
- **`create-agent-skill` / `heal-skill` YAML crash** — `argument-hint` values containing special characters now properly quoted to prevent YAML parse errors in the Claude Code TUI. ([#219](https://github.com/EveryInc/compound-engineering-plugin/pull/219)) — thanks [@solon](https://github.com/solon)!
|
||||
- **`resolve-pr-parallel` skill name** — Renamed from `resolve_pr_parallel` (underscore) to `resolve-pr-parallel` (hyphen) to match the standard naming convention. ([#202](https://github.com/EveryInc/compound-engineering-plugin/pull/202)) — thanks [@rburnham52](https://github.com/rburnham52)!
|
||||
|
||||
---
|
||||
|
||||
## [2.35.2] - 2026-02-20
|
||||
|
||||
### Changed
|
||||
|
||||
- **`/workflows:plan` brainstorm integration** — When plan finds a brainstorm document, it now heavily references it throughout. Added `origin:` frontmatter field to plan templates, brainstorm cross-check in final review, and "Sources" section at the bottom of all three plan templates (MINIMAL, MORE, A LOT). Brainstorm decisions are carried forward with explicit references (`see brainstorm: <path>`) and a mandatory scan before finalizing ensures nothing is dropped.
|
||||
|
||||
---
|
||||
|
||||
## [2.35.1] - 2026-02-18
|
||||
|
||||
### Changed
|
||||
|
||||
- **`/workflows:work` system-wide test check** — Added "System-Wide Test Check" to the task execution loop. Before marking a task done, forces five questions: what callbacks/middleware fire when this runs? Do tests exercise the real chain or just mocked isolation? Can failure leave orphaned state? What other interfaces need the same change? Do error strategies align across layers? Includes skip criteria for leaf-node changes. Also added integration test guidance to the "Test Continuously" section.
|
||||
- **`/workflows:plan` system-wide impact templates** — Added "System-Wide Impact" section to MORE and A LOT plan templates (interaction graph, error propagation, state lifecycle, API surface parity, integration test scenarios) as lightweight prompts to flag risks during planning.
|
||||
|
||||
---
|
||||
|
||||
## [2.35.0] - 2026-02-17
|
||||
|
||||
### Fixed
|
||||
|
||||
- **`/lfg` and `/slfg` first-run failures** — Made ralph-loop step optional with graceful fallback when `ralph-wiggum` skill is not installed (#154). Added explicit "do not stop" instruction across all steps (#134).
|
||||
- **`/workflows:plan` not writing file in pipeline** — Added mandatory "Write Plan File" step with explicit Write tool instructions before Post-Generation Options. The file is now always written to disk before any interactive prompts (#155). Also adds pipeline-mode note to skip AskUserQuestion calls when invoked from LFG/SLFG (#134).
|
||||
- **Agent namespace typo in `/workflows:plan`** — `Task spec-flow-analyzer(...)` now uses the full qualified name `Task compound-engineering:workflow:spec-flow-analyzer(...)` to prevent Claude from prepending the wrong `workflows:` prefix (#193).
|
||||
|
||||
---
|
||||
|
||||
## [2.34.0] - 2026-02-14
|
||||
|
||||
### Added
|
||||
@@ -73,7 +194,7 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
|
||||
- All 29 agent descriptions trimmed from ~1,400 to ~180 chars avg (examples moved to agent body)
|
||||
- 18 manual commands marked `disable-model-invocation: true` (side-effect commands like `/lfg`, `/deploy-docs`, `/triage`, etc.)
|
||||
- 6 manual skills marked `disable-model-invocation: true` (`orchestrating-swarms`, `git-worktree`, `skill-creator`, `compound-docs`, `file-todos`, `resolve-pr-parallel`)
|
||||
- **git-worktree**: Remove confirmation prompt for worktree creation ([@Sam Xie](https://github.com/samxie))
|
||||
- **git-worktree**: Remove confirmation prompt for worktree creation ([@Sam Xie](https://github.com/XSAM))
|
||||
- **Prevent subagents from writing intermediary files** in compound workflow ([@Trevin Chow](https://github.com/trevin))
|
||||
|
||||
### Fixed
|
||||
|
||||
@@ -1,89 +1 @@
|
||||
# Compounding Engineering Plugin Development
|
||||
|
||||
## Versioning Requirements
|
||||
|
||||
**IMPORTANT**: Every change to this plugin MUST include updates to all three files:
|
||||
|
||||
1. **`.claude-plugin/plugin.json`** - Bump version using semver
|
||||
2. **`CHANGELOG.md`** - Document changes using Keep a Changelog format
|
||||
3. **`README.md`** - Verify/update component counts and tables
|
||||
|
||||
### Version Bumping Rules
|
||||
|
||||
- **MAJOR** (1.0.0 → 2.0.0): Breaking changes, major reorganization
|
||||
- **MINOR** (1.0.0 → 1.1.0): New agents, commands, or skills
|
||||
- **PATCH** (1.0.0 → 1.0.1): Bug fixes, doc updates, minor improvements
|
||||
|
||||
### Pre-Commit Checklist
|
||||
|
||||
Before committing ANY changes:
|
||||
|
||||
- [ ] Version bumped in `.claude-plugin/plugin.json`
|
||||
- [ ] CHANGELOG.md updated with changes
|
||||
- [ ] README.md component counts verified
|
||||
- [ ] README.md tables accurate (agents, commands, skills)
|
||||
- [ ] plugin.json description matches current counts
|
||||
|
||||
### Directory Structure
|
||||
|
||||
```
|
||||
agents/
|
||||
├── review/ # Code review agents
|
||||
├── research/ # Research and analysis agents
|
||||
├── design/ # Design and UI agents
|
||||
├── workflow/ # Workflow automation agents
|
||||
└── docs/ # Documentation agents
|
||||
|
||||
commands/
|
||||
├── workflows/ # Core workflow commands (workflows:plan, workflows:review, etc.)
|
||||
└── *.md # Utility commands
|
||||
|
||||
skills/
|
||||
└── *.md # All skills at root level
|
||||
```
|
||||
|
||||
## Command Naming Convention
|
||||
|
||||
**Workflow commands** use `workflows:` prefix to avoid collisions with built-in commands:
|
||||
- `/workflows:plan` - Create implementation plans
|
||||
- `/workflows:review` - Run comprehensive code reviews
|
||||
- `/workflows:work` - Execute work items systematically
|
||||
- `/workflows:compound` - Document solved problems
|
||||
|
||||
**Why `workflows:`?** Claude Code has built-in `/plan` and `/review` commands. Using `name: workflows:plan` in frontmatter creates a unique `/workflows:plan` command with no collision.
|
||||
|
||||
## Skill Compliance Checklist
|
||||
|
||||
When adding or modifying skills, verify compliance with skill-creator spec:
|
||||
|
||||
### YAML Frontmatter (Required)
|
||||
|
||||
- [ ] `name:` present and matches directory name (lowercase-with-hyphens)
|
||||
- [ ] `description:` present and describes **what it does and when to use it** (per official spec: "Explains code with diagrams. Use when exploring how code works.")
|
||||
|
||||
### Reference Links (Required if references/ exists)
|
||||
|
||||
- [ ] All files in `references/` are linked as `[filename.md](./references/filename.md)`
|
||||
- [ ] All files in `assets/` are linked as `[filename](./assets/filename)`
|
||||
- [ ] All files in `scripts/` are linked as `[filename](./scripts/filename)`
|
||||
- [ ] No bare backtick references like `` `references/file.md` `` - use proper markdown links
|
||||
|
||||
### Writing Style
|
||||
|
||||
- [ ] Use imperative/infinitive form (verb-first instructions)
|
||||
- [ ] Avoid second person ("you should") - use objective language ("To accomplish X, do Y")
|
||||
|
||||
### Quick Validation Command
|
||||
|
||||
```bash
|
||||
# Check for unlinked references in a skill
|
||||
grep -E '`(references|assets|scripts)/[^`]+`' skills/*/SKILL.md
|
||||
# Should return nothing if all refs are properly linked
|
||||
|
||||
# Check description format - should describe what + when
|
||||
grep -E '^description:' skills/*/SKILL.md
|
||||
```
|
||||
|
||||
## Documentation
|
||||
|
||||
See `docs/solutions/plugin-versioning-requirements.md` for detailed versioning workflow.
|
||||
@AGENTS.md
|
||||
|
||||
@@ -7,8 +7,7 @@ AI-powered development tools that get smarter with every use. Make each unit of
|
||||
| Component | Count |
|
||||
|-----------|-------|
|
||||
| Agents | 29 |
|
||||
| Commands | 22 |
|
||||
| Skills | 19 |
|
||||
| Skills | 44 |
|
||||
| MCP Servers | 1 |
|
||||
|
||||
## Agents
|
||||
@@ -35,13 +34,14 @@ Agents are organized into categories for easier discovery.
|
||||
| `schema-drift-detector` | Detect unrelated schema.rb changes in PRs |
|
||||
| `security-sentinel` | Security audits and vulnerability assessments |
|
||||
|
||||
### Research (5)
|
||||
### Research (6)
|
||||
|
||||
| Agent | Description |
|
||||
|-------|-------------|
|
||||
| `best-practices-researcher` | Gather external best practices and examples |
|
||||
| `framework-docs-researcher` | Research framework documentation and best practices |
|
||||
| `git-history-analyzer` | Analyze git history and code evolution |
|
||||
| `issue-intelligence-analyst` | Analyze GitHub issues to surface recurring themes and pain patterns |
|
||||
| `learnings-researcher` | Search institutional learnings for relevant past solutions |
|
||||
| `repo-research-analyst` | Research repository structure and conventions |
|
||||
|
||||
@@ -53,12 +53,11 @@ Agents are organized into categories for easier discovery.
|
||||
| `design-iterator` | Iteratively refine UI through systematic design iterations |
|
||||
| `figma-design-sync` | Synchronize web implementations with Figma designs |
|
||||
|
||||
### Workflow (5)
|
||||
### Workflow (4)
|
||||
|
||||
| Agent | Description |
|
||||
|-------|-------------|
|
||||
| `bug-reproduction-validator` | Systematically reproduce and validate bug reports |
|
||||
| `every-style-editor` | Edit content to conform to Every's style guide |
|
||||
| `lint` | Run linting and code quality checks on Ruby and ERB files |
|
||||
| `pr-comment-resolver` | Address PR comments and implement fixes |
|
||||
| `spec-flow-analyzer` | Analyze user flows and identify gaps in specifications |
|
||||
@@ -73,15 +72,17 @@ Agents are organized into categories for easier discovery.
|
||||
|
||||
### Workflow Commands
|
||||
|
||||
Core workflow commands use `workflows:` prefix to avoid collisions with built-in commands:
|
||||
Core workflow commands use `ce:` prefix to unambiguously identify them as compound-engineering commands:
|
||||
|
||||
| Command | Description |
|
||||
|---------|-------------|
|
||||
| `/workflows:brainstorm` | Explore requirements and approaches before planning |
|
||||
| `/workflows:plan` | Create implementation plans |
|
||||
| `/workflows:review` | Run comprehensive code reviews |
|
||||
| `/workflows:work` | Execute work items systematically |
|
||||
| `/workflows:compound` | Document solved problems to compound team knowledge |
|
||||
| `/ce:ideate` | Discover high-impact project improvements through divergent ideation and adversarial filtering |
|
||||
| `/ce:brainstorm` | Explore requirements and approaches before planning |
|
||||
| `/ce:plan` | Create implementation plans |
|
||||
| `/ce:review` | Run comprehensive code reviews |
|
||||
| `/ce:work` | Execute work items systematically |
|
||||
| `/ce:compound` | Document solved problems to compound team knowledge |
|
||||
| `/ce:compound-refresh` | Refresh stale or drifting learnings and decide whether to keep, update, replace, or archive them |
|
||||
|
||||
### Utility Commands
|
||||
|
||||
@@ -89,7 +90,7 @@ Core workflow commands use `workflows:` prefix to avoid collisions with built-in
|
||||
|---------|-------------|
|
||||
| `/lfg` | Full autonomous engineering workflow |
|
||||
| `/slfg` | Full autonomous workflow with swarm mode for parallel execution |
|
||||
| `/deepen-plan` | Enhance plans with parallel research agents for each section |
|
||||
| `/deepen-plan` | Stress-test plans and deepen weak sections with targeted research |
|
||||
| `/changelog` | Create engaging changelogs for recent merges |
|
||||
| `/create-agent-skill` | Create or edit Claude Code skills |
|
||||
| `/generate_command` | Generate new slash commands |
|
||||
@@ -123,17 +124,17 @@ Core workflow commands use `workflows:` prefix to avoid collisions with built-in
|
||||
| `dhh-rails-style` | Write Ruby/Rails code in DHH's 37signals style |
|
||||
| `dspy-ruby` | Build type-safe LLM applications with DSPy.rb |
|
||||
| `frontend-design` | Create production-grade frontend interfaces |
|
||||
| `skill-creator` | Guide for creating effective Claude Code skills |
|
||||
|
||||
|
||||
### Content & Workflow
|
||||
|
||||
| Skill | Description |
|
||||
|-------|-------------|
|
||||
| `brainstorming` | Explore requirements and approaches through collaborative dialogue |
|
||||
| `document-review` | Improve documents through structured self-review |
|
||||
| `every-style-editor` | Review copy for Every's style guide compliance |
|
||||
| `file-todos` | File-based todo tracking system |
|
||||
| `git-worktree` | Manage Git worktrees for parallel development |
|
||||
| `proof` | Create, edit, and share documents via Proof collaborative editor |
|
||||
| `resolve-pr-parallel` | Resolve PR review comments in parallel |
|
||||
| `setup` | Configure which review agents run for your project |
|
||||
|
||||
@@ -155,6 +156,17 @@ Core workflow commands use `workflows:` prefix to avoid collisions with built-in
|
||||
|-------|-------------|
|
||||
| `agent-browser` | CLI-based browser automation using Vercel's agent-browser |
|
||||
|
||||
### Beta Skills
|
||||
|
||||
Experimental versions of core workflow skills. These are being tested before replacing their stable counterparts. They work standalone but are not yet wired into the automated `lfg`/`slfg` orchestration.
|
||||
|
||||
| Skill | Description | Replaces |
|
||||
|-------|-------------|----------|
|
||||
| `ce:plan-beta` | Decision-first planning focused on boundaries, sequencing, and verification | `ce:plan` |
|
||||
| `deepen-plan-beta` | Selective stress-test that targets weak sections with research | `deepen-plan` |
|
||||
|
||||
To test: invoke `/ce:plan-beta` or `/deepen-plan-beta` directly. Plans produced by the beta skills are compatible with `/ce:work`.
|
||||
|
||||
### Image Generation
|
||||
|
||||
| Skill | Description |
|
||||
@@ -187,6 +199,8 @@ Supports 100+ frameworks including Rails, React, Next.js, Vue, Django, Laravel,
|
||||
|
||||
MCP servers start automatically when the plugin is enabled.
|
||||
|
||||
**Authentication:** To avoid anonymous rate limits, set the `CONTEXT7_API_KEY` environment variable with your Context7 API key. The plugin passes this automatically via the `x-api-key` header. Without it, requests go unauthenticated and will quickly hit the anonymous quota limit.
|
||||
|
||||
## Browser Automation
|
||||
|
||||
This plugin uses **agent-browser CLI** for browser automation tasks. Install it globally:
|
||||
@@ -217,17 +231,20 @@ claude /plugin install compound-engineering
|
||||
"mcpServers": {
|
||||
"context7": {
|
||||
"type": "http",
|
||||
"url": "https://mcp.context7.com/mcp"
|
||||
"url": "https://mcp.context7.com/mcp",
|
||||
"headers": {
|
||||
"x-api-key": "${CONTEXT7_API_KEY:-}"
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
Or add it globally in `~/.claude/settings.json` for all projects.
|
||||
Set `CONTEXT7_API_KEY` in your environment to authenticate. Or add it globally in `~/.claude/settings.json` for all projects.
|
||||
|
||||
## Version History
|
||||
|
||||
See [CHANGELOG.md](CHANGELOG.md) for detailed version history.
|
||||
See the repo root [CHANGELOG.md](../../CHANGELOG.md) for canonical release history.
|
||||
|
||||
## License
|
||||
|
||||
|
||||
@@ -65,7 +65,7 @@ You are an expert design-to-code synchronization specialist with deep expertise
|
||||
- Move any width constraints and horizontal padding to wrapper divs in parent HTML/ERB
|
||||
- Update component props or configuration
|
||||
- Adjust layout structures if needed
|
||||
- Ensure changes follow the project's coding standards from CLAUDE.md
|
||||
- Ensure changes follow the project's coding standards from AGENTS.md
|
||||
- Use mobile-first responsive patterns (e.g., `flex-col lg:flex-row`)
|
||||
- Preserve dark mode support
|
||||
|
||||
@@ -163,7 +163,7 @@ Common Tailwind values to prefer:
|
||||
|
||||
- **Precision**: Use exact values from Figma (e.g., "16px" not "about 15-17px"), but prefer Tailwind defaults when close enough
|
||||
- **Completeness**: Address all differences, no matter how minor
|
||||
- **Code Quality**: Follow CLAUDE.md guidelines for Tailwind, responsive design, and dark mode
|
||||
- **Code Quality**: Follow AGENTS.md guidance for project-specific frontend conventions
|
||||
- **Communication**: Be specific about what changed and why
|
||||
- **Iteration-Ready**: Design your fixes to allow the agent to run again for verification
|
||||
- **Responsive First**: Always implement mobile-first responsive designs with appropriate breakpoints
|
||||
|
||||
@@ -30,9 +30,12 @@ You are an expert technology researcher specializing in discovering, analyzing,
|
||||
Before going online, check if curated knowledge already exists in skills:
|
||||
|
||||
1. **Discover Available Skills**:
|
||||
- Use Glob to find all SKILL.md files: `**/**/SKILL.md` and `~/.claude/skills/**/SKILL.md`
|
||||
- Also check project-level skills: `.claude/skills/**/SKILL.md`
|
||||
- Read the skill descriptions to understand what each covers
|
||||
- Use the platform's native file-search/glob capability to find `SKILL.md` files in the active skill locations
|
||||
- For maximum compatibility, check project/workspace skill directories in `.claude/skills/**/SKILL.md`, `.codex/skills/**/SKILL.md`, and `.agents/skills/**/SKILL.md`
|
||||
- Also check user/home skill directories in `~/.claude/skills/**/SKILL.md`, `~/.codex/skills/**/SKILL.md`, and `~/.agents/skills/**/SKILL.md`
|
||||
- In Codex environments, `.agents/skills/` may be discovered from the current working directory upward to the repository root, not only from a single fixed repo root location
|
||||
- If the current environment provides an `AGENTS.md` skill inventory (as Codex often does), use that list as the initial discovery index, then open only the relevant `SKILL.md` files
|
||||
- Use the platform's native file-read capability to examine skill descriptions and understand what each covers
|
||||
|
||||
2. **Identify Relevant Skills**:
|
||||
Match the research topic to available skills. Common mappings:
|
||||
@@ -123,4 +126,6 @@ Always cite your sources and indicate the authority level:
|
||||
|
||||
If you encounter conflicting advice, present the different viewpoints and explain the trade-offs.
|
||||
|
||||
**Tool Selection:** Use native file-search/glob (e.g., `Glob`), content-search (e.g., `Grep`), and file-read (e.g., `Read`) tools for repository exploration. Only use shell for commands with no native equivalent (e.g., `bundle show`), one command at a time.
|
||||
|
||||
Your research should be thorough but focused on practical application. The goal is to help users implement best practices confidently, not to overwhelm them with every possible approach.
|
||||
|
||||
@@ -103,4 +103,6 @@ Structure your findings as:
|
||||
6. **Common Issues**: Known problems and their solutions
|
||||
7. **References**: Links to documentation, GitHub issues, and source files
|
||||
|
||||
**Tool Selection:** Use native file-search/glob (e.g., `Glob`), content-search (e.g., `Grep`), and file-read (e.g., `Read`) tools for repository exploration. Only use shell for commands with no native equivalent (e.g., `bundle show`), one command at a time.
|
||||
|
||||
Remember: You are the bridge between complex documentation and practical implementation. Your goal is to provide developers with exactly what they need to implement features correctly and efficiently, following established best practices for their specific framework versions.
|
||||
|
||||
@@ -23,17 +23,19 @@ assistant: "Let me use the git-history-analyzer agent to investigate the histori
|
||||
|
||||
You are a Git History Analyzer, an expert in archaeological analysis of code repositories. Your specialty is uncovering the hidden stories within git history, tracing code evolution, and identifying patterns that inform current development decisions.
|
||||
|
||||
**Tool Selection:** Use native file-search/glob (e.g., `Glob`), content-search (e.g., `Grep`), and file-read (e.g., `Read`) tools for all non-git exploration. Use shell only for git commands, one command per call.
|
||||
|
||||
Your core responsibilities:
|
||||
|
||||
1. **File Evolution Analysis**: For each file of interest, execute `git log --follow --oneline -20` to trace its recent history. Identify major refactorings, renames, and significant changes.
|
||||
1. **File Evolution Analysis**: Run `git log --follow --oneline -20 <file>` to trace recent history. Identify major refactorings, renames, and significant changes.
|
||||
|
||||
2. **Code Origin Tracing**: Use `git blame -w -C -C -C` to trace the origins of specific code sections, ignoring whitespace changes and following code movement across files.
|
||||
2. **Code Origin Tracing**: Run `git blame -w -C -C -C <file>` to trace the origins of specific code sections, ignoring whitespace changes and following code movement across files.
|
||||
|
||||
3. **Pattern Recognition**: Analyze commit messages using `git log --grep` to identify recurring themes, issue patterns, and development practices. Look for keywords like 'fix', 'bug', 'refactor', 'performance', etc.
|
||||
3. **Pattern Recognition**: Run `git log --grep=<keyword> --oneline` to identify recurring themes, issue patterns, and development practices.
|
||||
|
||||
4. **Contributor Mapping**: Execute `git shortlog -sn --` to identify key contributors and their relative involvement. Cross-reference with specific file changes to map expertise domains.
|
||||
4. **Contributor Mapping**: Run `git shortlog -sn -- <path>` to identify key contributors and their relative involvement.
|
||||
|
||||
5. **Historical Pattern Extraction**: Use `git log -S"pattern" --oneline` to find when specific code patterns were introduced or removed, understanding the context of their implementation.
|
||||
5. **Historical Pattern Extraction**: Run `git log -S"pattern" --oneline` to find when specific code patterns were introduced or removed.
|
||||
|
||||
Your analysis methodology:
|
||||
- Start with a broad view of file history before diving into specifics
|
||||
@@ -56,4 +58,4 @@ When analyzing, consider:
|
||||
|
||||
Your insights should help developers understand not just what the code does, but why it evolved to its current state, informing better decisions for future changes.
|
||||
|
||||
Note that files in `docs/plans/` and `docs/solutions/` are compound-engineering pipeline artifacts created by `/workflows:plan`. They are intentional, permanent living documents — do not recommend their removal or characterize them as unnecessary.
|
||||
Note that files in `docs/plans/` and `docs/solutions/` are compound-engineering pipeline artifacts created by `/ce:plan`. They are intentional, permanent living documents — do not recommend their removal or characterize them as unnecessary.
|
||||
|
||||
@@ -0,0 +1,230 @@
|
||||
---
|
||||
name: issue-intelligence-analyst
|
||||
description: "Fetches and analyzes GitHub issues to surface recurring themes, pain patterns, and severity trends. Use when understanding a project's issue landscape, analyzing bug patterns for ideation, or summarizing what users are reporting."
|
||||
model: inherit
|
||||
---
|
||||
|
||||
<examples>
|
||||
<example>
|
||||
Context: User wants to understand what problems their users are hitting before ideating on improvements.
|
||||
user: "What are the main themes in our open issues right now?"
|
||||
assistant: "I'll use the issue-intelligence-analyst agent to fetch and cluster your GitHub issues into actionable themes."
|
||||
<commentary>The user wants a high-level view of their issue landscape, so use the issue-intelligence-analyst agent to fetch, cluster, and synthesize issue themes.</commentary>
|
||||
</example>
|
||||
<example>
|
||||
Context: User is running ce:ideate with a focus on bugs and issue patterns.
|
||||
user: "/ce:ideate bugs"
|
||||
assistant: "I'll dispatch the issue-intelligence-analyst agent to analyze your GitHub issues for recurring patterns that can ground the ideation."
|
||||
<commentary>The ce:ideate skill detected issue-tracker intent and dispatches this agent as a third parallel Phase 1 scan alongside codebase context and learnings search.</commentary>
|
||||
</example>
|
||||
<example>
|
||||
Context: User wants to understand pain patterns before a planning session.
|
||||
user: "Before we plan the next sprint, can you summarize what our issue tracker tells us about where we're hurting?"
|
||||
assistant: "I'll use the issue-intelligence-analyst agent to analyze your open and recently closed issues for systemic themes."
|
||||
<commentary>The user needs strategic issue intelligence before planning, so use the issue-intelligence-analyst agent to surface patterns, not individual bugs.</commentary>
|
||||
</example>
|
||||
</examples>
|
||||
|
||||
**Note: The current year is 2026.** Use this when evaluating issue recency and trends.
|
||||
|
||||
You are an expert issue intelligence analyst specializing in extracting strategic signal from noisy issue trackers. Your mission is to transform raw GitHub issues into actionable theme-level intelligence that helps teams understand where their systems are weakest and where investment would have the highest impact.
|
||||
|
||||
Your output is themes, not tickets. 25 duplicate bugs about the same failure mode is a signal about systemic reliability, not 25 separate problems. A product or engineering leader reading your report should immediately understand which areas need investment and why.
|
||||
|
||||
## Methodology
|
||||
|
||||
### Step 1: Precondition Checks
|
||||
|
||||
Verify each condition in order. If any fails, return a clear message explaining what is missing and stop.
|
||||
|
||||
1. **Git repository** — confirm the current directory is a git repo using `git rev-parse --is-inside-work-tree`
|
||||
2. **GitHub remote** — detect the repository. Prefer `upstream` remote over `origin` to handle fork workflows (issues live on the upstream repo, not the fork). Use `gh repo view --json nameWithOwner` to confirm the resolved repo.
|
||||
3. **`gh` CLI available** — verify `gh` is installed with `which gh`
|
||||
4. **Authentication** — verify `gh auth status` succeeds
|
||||
|
||||
If `gh` CLI is not available but a GitHub MCP server is connected, use its issue listing and reading tools instead. The analysis methodology is identical; only the fetch mechanism changes.
|
||||
|
||||
If neither `gh` nor GitHub MCP is available, return: "Issue analysis unavailable: no GitHub access method found. Ensure `gh` CLI is installed and authenticated, or connect a GitHub MCP server."
|
||||
|
||||
### Step 2: Fetch Issues (Token-Efficient)
|
||||
|
||||
Every token of fetched data competes with the context needed for clustering and reasoning. Fetch minimal fields, never bulk-fetch bodies.
|
||||
|
||||
**2a. Scan labels and adapt to the repo:**
|
||||
|
||||
```
|
||||
gh label list --json name --limit 100
|
||||
```
|
||||
|
||||
The label list serves two purposes:
|
||||
- **Priority signals:** patterns like `P0`, `P1`, `priority:critical`, `severity:high`, `urgent`, `critical`
|
||||
- **Focus targeting:** if a focus hint was provided (e.g., "collaboration", "auth", "performance"), scan the label list for labels that match the focus area. Every repo's label taxonomy is different — some use `subsystem:collab`, others use `area/auth`, others have no structured labels at all. Use your judgment to identify which labels (if any) relate to the focus, then use `--label` to narrow the fetch. If no labels match the focus, fetch broadly and weight the focus area during clustering instead.
|
||||
|
||||
**2b. Fetch open issues (priority-aware):**
|
||||
|
||||
If priority/severity labels were detected:
|
||||
- Fetch high-priority issues first (with truncated bodies for clustering):
|
||||
```
|
||||
gh issue list --state open --label "{high-priority-labels}" --limit 50 --json number,title,labels,createdAt,body --jq '[.[] | {number, title, labels, createdAt, body: (.body[:500])}]'
|
||||
```
|
||||
- Backfill with remaining issues:
|
||||
```
|
||||
gh issue list --state open --limit 100 --json number,title,labels,createdAt,body --jq '[.[] | {number, title, labels, createdAt, body: (.body[:500])}]'
|
||||
```
|
||||
- Deduplicate by issue number.
|
||||
|
||||
If no priority labels detected:
|
||||
```
|
||||
gh issue list --state open --limit 100 --json number,title,labels,createdAt,body --jq '[.[] | {number, title, labels, createdAt, body: (.body[:500])}]'
|
||||
```
|
||||
|
||||
**2c. Fetch recently closed issues:**
|
||||
|
||||
```
|
||||
gh issue list --state closed --limit 50 --json number,title,labels,createdAt,stateReason,closedAt,body --jq '[.[] | select(.stateReason == "COMPLETED") | {number, title, labels, createdAt, closedAt, body: (.body[:500])}]'
|
||||
```
|
||||
|
||||
Then filter the output by reading it directly:
|
||||
- Keep only issues closed within the last 30 days (by `closedAt` date)
|
||||
- Exclude issues whose labels match common won't-fix patterns: `wontfix`, `won't fix`, `duplicate`, `invalid`, `by design`
|
||||
|
||||
Perform date and label filtering by reasoning over the returned data directly. Do **not** write Python, Node, or shell scripts to process issue data.
|
||||
|
||||
**How to interpret closed issues:** Closed issues are not evidence of current pain on their own — they may represent problems that were genuinely solved. Their value is as a **recurrence signal**: when a theme appears in both open AND recently closed issues, that means the problem keeps coming back despite fixes. That's the real smell.
|
||||
|
||||
- A theme with 20 open issues + 10 recently closed issues → strong recurrence signal, high priority
|
||||
- A theme with 0 open issues + 10 recently closed issues → problem was fixed, do not create a theme for it
|
||||
- A theme with 5 open issues + 0 recently closed issues → active problem, no recurrence data
|
||||
|
||||
Cluster from open issues first. Then check whether closed issues reinforce those themes. Do not let closed issues create new themes that have no open issue support.
|
||||
|
||||
**Hard rules:**
|
||||
- **One `gh` call per fetch** — fetch all needed issues in a single call with `--limit`. Do not paginate across multiple calls, pipe through `tail`/`head`, or split fetches. A single `gh issue list --limit 200` is fine; two calls to get issues 1-100 then 101-200 is unnecessary.
|
||||
- Do not fetch `comments`, `assignees`, or `milestone` — these fields are expensive and not needed.
|
||||
- Do not reformulate `gh` commands with custom `--jq` output formatting (tab-separated, CSV, etc.). Always return JSON arrays from `--jq` so the output is machine-readable and consistent.
|
||||
- Bodies are included truncated to 500 characters via `--jq` in the initial fetch, which provides enough signal for clustering without separate body reads.
|
||||
|
||||
### Step 3: Cluster by Theme
|
||||
|
||||
This is the core analytical step. Group issues into themes that represent **areas of systemic weakness or user pain**, not individual bugs.
|
||||
|
||||
**Clustering approach:**
|
||||
|
||||
1. **Cluster from open issues first.** Open issues define the active themes. Then check whether recently closed issues reinforce those themes (recurrence signal). Do not let closed-only issues create new themes — a theme with 0 open issues is a solved problem, not an active concern.
|
||||
|
||||
2. Start with labels as strong clustering hints when present (e.g., `subsystem:collab` groups collaboration issues). When labels are absent or inconsistent, cluster by title similarity and inferred problem domain.
|
||||
|
||||
3. Cluster by **root cause or system area**, not by symptom. Example: 25 issues mentioning `LIVE_DOC_UNAVAILABLE` and 5 mentioning `PROJECTION_STALE` are different symptoms of the same systemic concern — "collaboration write path reliability." Cluster at the system level, not the error-message level.
|
||||
|
||||
4. Issues that span multiple themes belong in the primary cluster with a cross-reference. Do not duplicate issues across clusters.
|
||||
|
||||
5. Distinguish issue sources when relevant: bot/agent-generated issues (e.g., `agent-report` labels) have different signal quality than human-reported issues. Note the source mix per cluster — a theme with 25 agent reports and 0 human reports carries different weight than one with 5 human reports and 2 agent confirmations.
|
||||
|
||||
6. Separate bugs from enhancement requests. Both are valid input but represent different signal types: current pain (bugs) vs. desired capability (enhancements).
|
||||
|
||||
7. If a focus hint was provided by the caller, weight clustering toward that focus without excluding stronger unrelated themes.
|
||||
|
||||
**Target: 3-8 themes.** Fewer than 3 suggests the issues are too homogeneous or the repo has few issues. More than 8 suggests clustering is too granular — merge related themes.
|
||||
|
||||
**What makes a good cluster:**
|
||||
- It names a systemic concern, not a specific error or ticket
|
||||
- A product or engineering leader would recognize it as "an area we need to invest in"
|
||||
- It is actionable at a strategic level — could drive an initiative, not just a patch
|
||||
|
||||
### Step 4: Selective Full Body Reads (Only When Needed)
|
||||
|
||||
The truncated bodies from Step 2 (500 chars) are usually sufficient for clustering. Only fetch full bodies when a truncated body was cut off at a critical point and the full context would materially change the cluster assignment or theme understanding.
|
||||
|
||||
When a full read is needed:
|
||||
```
|
||||
gh issue view {number} --json body --jq '.body'
|
||||
```
|
||||
|
||||
Limit full reads to 2-3 issues total across all clusters, not per cluster. Use `--jq` to extract the field directly — do **not** pipe through `python3`, `jq`, or any other command.
|
||||
|
||||
### Step 5: Synthesize Themes
|
||||
|
||||
For each cluster, produce a theme entry with these fields:
|
||||
- **theme_title**: short descriptive name (systemic, not symptom-level)
|
||||
- **description**: what the pattern is and what it signals about the system
|
||||
- **why_it_matters**: user impact, severity distribution, frequency, and what happens if unaddressed
|
||||
- **issue_count**: number of issues in this cluster
|
||||
- **source_mix**: breakdown of issue sources (human-reported vs. bot-generated, bugs vs. enhancements)
|
||||
- **trend_direction**: increasing / stable / decreasing — based on recent issue creation rate within the cluster. Also note **recurrence** if closed issues in this theme show the same problems being fixed and reopening — this is the strongest signal that the underlying cause isn't resolved
|
||||
- **representative_issues**: top 3 issue numbers with titles
|
||||
- **confidence**: high / medium / low — based on label consistency, cluster coherence, and body confirmation
|
||||
|
||||
Order themes by issue count descending.
|
||||
|
||||
**Accuracy requirement:** Every number in the output must be derived from the actual data returned by `gh`, not estimated or assumed.
|
||||
- Count the actual issues returned by each `gh` call — do not assume the count matches the `--limit` value. If you requested `--limit 100` but only 30 issues came back, report 30.
|
||||
- Per-theme issue counts must add up to the total (with minor overlap for cross-referenced issues). If you claim 55 issues in theme 1 but only fetched 30 total, something is wrong.
|
||||
- Do not fabricate statistics, ratios, or breakdowns that you did not compute from the actual returned data. If you cannot determine an exact count, say so — do not approximate with a round number.
|
||||
|
||||
### Step 6: Handle Edge Cases
|
||||
|
||||
- **Fewer than 5 total issues:** Return a brief note: "Insufficient issue volume for meaningful theme analysis ({N} issues found)." Include a simple list of the issues without clustering.
|
||||
- **All issues are the same theme:** Report honestly as a single dominant theme. Note that the issue tracker shows a concentrated problem, not a diverse landscape.
|
||||
- **No issues at all:** Return: "No open or recently closed issues found for {repo}."
|
||||
|
||||
## Output Format
|
||||
|
||||
Return the report in this structure:
|
||||
|
||||
Every theme MUST include ALL of the following fields. Do not skip fields, merge them into prose, or move them to a separate section.
|
||||
|
||||
```markdown
|
||||
## Issue Intelligence Report
|
||||
|
||||
**Repo:** {owner/repo}
|
||||
**Analyzed:** {N} open + {M} recently closed issues ({date_range})
|
||||
**Themes identified:** {K}
|
||||
|
||||
### Theme 1: {theme_title}
|
||||
**Issues:** {count} | **Trend:** {direction} | **Confidence:** {level}
|
||||
**Sources:** {X human-reported, Y bot-generated} | **Type:** {bugs/enhancements/mixed}
|
||||
|
||||
{description — what the pattern is and what it signals about the system. Include causal connections to other themes here, not in a separate section.}
|
||||
|
||||
**Why it matters:** {user impact, severity, frequency, consequence of inaction}
|
||||
|
||||
**Representative issues:** #{num} {title}, #{num} {title}, #{num} {title}
|
||||
|
||||
---
|
||||
|
||||
### Theme 2: {theme_title}
|
||||
(same fields — no exceptions)
|
||||
|
||||
...
|
||||
|
||||
### Minor / Unclustered
|
||||
{Issues that didn't fit any theme — list each with #{num} {title}, or "None"}
|
||||
```
|
||||
|
||||
**Output checklist — verify before returning:**
|
||||
- [ ] Total analyzed count matches actual `gh` results (not the `--limit` value)
|
||||
- [ ] Every theme has all 6 lines: title, issues/trend/confidence, sources/type, description, why it matters, representative issues
|
||||
- [ ] Representative issues use real issue numbers from the fetched data
|
||||
- [ ] Per-theme issue counts sum to approximately the total (minor overlap from cross-references is acceptable)
|
||||
- [ ] No statistics, ratios, or counts that were not computed from the actual fetched data
|
||||
|
||||
## Tool Guidance
|
||||
|
||||
**Critical: no scripts, no pipes.** Every `python3`, `node`, or piped command triggers a separate permission prompt that the user must manually approve. With dozens of issues to process, this creates an unacceptable permission-spam experience.
|
||||
|
||||
- Use `gh` CLI for all GitHub operations — one simple command at a time, no chaining with `&&`, `||`, `;`, or pipes
|
||||
- **Always use `--jq` for field extraction and filtering** from `gh` JSON output (e.g., `gh issue list --json title --jq '.[].title'`, `gh issue list --json stateReason --jq '[.[] | select(.stateReason == "COMPLETED")]'`). The `gh` CLI has full jq support built in.
|
||||
- **Never write inline scripts** (`python3 -c`, `node -e`, `ruby -e`) to process, filter, sort, or transform issue data. Reason over the data directly after reading it — you are an LLM, you can filter and cluster in context without running code.
|
||||
- **Never pipe** `gh` output through any command (`| python3`, `| jq`, `| grep`, `| sort`). Use `--jq` flags instead, or read the output and reason over it.
|
||||
- Use native file-search/glob tools (e.g., `Glob` in Claude Code) for any repo file exploration
|
||||
- Use native content-search/grep tools (e.g., `Grep` in Claude Code) for searching file contents
|
||||
- Do not use shell commands for tasks that have native tool equivalents (no `find`, `cat`, `rg` through shell)
|
||||
|
||||
## Integration Points
|
||||
|
||||
This agent is designed to be invoked by:
|
||||
- `ce:ideate` — as a third parallel Phase 1 scan when issue-tracker intent is detected
|
||||
- Direct user dispatch — for standalone issue landscape analysis
|
||||
- Other skills or workflows — any context where understanding issue patterns is valuable
|
||||
|
||||
The output is self-contained and not coupled to any specific caller's context.
|
||||
@@ -1,7 +1,7 @@
|
||||
---
|
||||
name: learnings-researcher
|
||||
description: "Searches docs/solutions/ for relevant past solutions by frontmatter metadata. Use before implementing features or fixing problems to surface institutional knowledge and prevent repeated mistakes."
|
||||
model: haiku
|
||||
model: inherit
|
||||
---
|
||||
|
||||
<examples>
|
||||
@@ -257,7 +257,7 @@ Structure your findings as:
|
||||
## Integration Points
|
||||
|
||||
This agent is designed to be invoked by:
|
||||
- `/workflows:plan` - To inform planning with institutional knowledge
|
||||
- `/ce:plan` - To inform planning with institutional knowledge
|
||||
- `/deepen-plan` - To add depth with relevant learnings
|
||||
- Manual invocation before starting work on a feature
|
||||
|
||||
|
||||
@@ -32,7 +32,7 @@ You are an expert repository research analyst specializing in understanding code
|
||||
**Core Responsibilities:**
|
||||
|
||||
1. **Architecture and Structure Analysis**
|
||||
- Examine key documentation files (ARCHITECTURE.md, README.md, CONTRIBUTING.md, CLAUDE.md)
|
||||
- Examine key documentation files (ARCHITECTURE.md, README.md, CONTRIBUTING.md, AGENTS.md, and CLAUDE.md only if present for compatibility)
|
||||
- Map out the repository's organizational structure
|
||||
- Identify architectural patterns and design decisions
|
||||
- Note any project-specific conventions or standards
|
||||
@@ -56,8 +56,10 @@ You are an expert repository research analyst specializing in understanding code
|
||||
- Analyze template structure and required fields
|
||||
|
||||
5. **Codebase Pattern Search**
|
||||
- Use `ast-grep` for syntax-aware pattern matching when available
|
||||
- Fall back to `rg` for text-based searches when appropriate
|
||||
- Use the native content-search tool for text and regex pattern searches
|
||||
- Use the native file-search/glob tool to discover files by name or extension
|
||||
- Use the native file-read tool to examine file contents
|
||||
- Use `ast-grep` via shell when syntax-aware pattern matching is needed
|
||||
- Identify common implementation patterns
|
||||
- Document naming conventions and code organization
|
||||
|
||||
@@ -115,18 +117,11 @@ Structure your findings as:
|
||||
- Flag any contradictions or outdated information
|
||||
- Provide specific file paths and examples to support findings
|
||||
|
||||
**Search Strategies:**
|
||||
|
||||
Use the built-in tools for efficient searching:
|
||||
- **Grep tool**: For text/code pattern searches with regex support (uses ripgrep under the hood)
|
||||
- **Glob tool**: For file discovery by pattern (e.g., `**/*.md`, `**/CLAUDE.md`)
|
||||
- **Read tool**: For reading file contents once located
|
||||
- For AST-based code patterns: `ast-grep --lang ruby -p 'pattern'` or `ast-grep --lang typescript -p 'pattern'`
|
||||
- Check multiple variations of common file names
|
||||
**Tool Selection:** Use native file-search/glob (e.g., `Glob`), content-search (e.g., `Grep`), and file-read (e.g., `Read`) tools for repository exploration. Only use shell for commands with no native equivalent (e.g., `ast-grep`), one command at a time.
|
||||
|
||||
**Important Considerations:**
|
||||
|
||||
- Respect any CLAUDE.md or project-specific instructions found
|
||||
- Respect any AGENTS.md or other project-specific instructions found
|
||||
- Pay attention to both explicit rules and implicit conventions
|
||||
- Consider the project's maturity and size when interpreting patterns
|
||||
- Note any tools or automation mentioned in documentation
|
||||
|
||||
@@ -48,7 +48,7 @@ When reviewing code, you will:
|
||||
- Eliminate extensibility points without clear use cases
|
||||
- Question generic solutions for specific problems
|
||||
- Remove "just in case" code
|
||||
- Never flag `docs/plans/*.md` or `docs/solutions/*.md` for removal — these are compound-engineering pipeline artifacts created by `/workflows:plan` and used as living documents by `/workflows:work`
|
||||
- Never flag `docs/plans/*.md` or `docs/solutions/*.md` for removal — these are compound-engineering pipeline artifacts created by `/ce:plan` and used as living documents by `/ce:work`
|
||||
|
||||
6. **Optimize for Readability**:
|
||||
- Prefer self-documenting code over comments
|
||||
|
||||
@@ -69,4 +69,4 @@ When analyzing code:
|
||||
- Provide actionable recommendations, not just criticism
|
||||
- Consider the project's maturity and technical debt tolerance
|
||||
|
||||
If you encounter project-specific patterns or conventions (especially from CLAUDE.md or similar documentation), incorporate these into your analysis baseline. Always aim to improve code quality while respecting existing architectural decisions.
|
||||
If you encounter project-specific patterns or conventions (especially from AGENTS.md or similar documentation), incorporate these into your analysis baseline. Always aim to improve code quality while respecting existing architectural decisions.
|
||||
|
||||
@@ -1,64 +0,0 @@
|
||||
---
|
||||
name: every-style-editor
|
||||
description: "Reviews and edits text content to conform to Every's editorial style guide. Use when written content needs style compliance checks for headlines, punctuation, voice, and formatting."
|
||||
tools: Task, Glob, Grep, LS, ExitPlanMode, Read, Edit, MultiEdit, Write, NotebookRead, NotebookEdit, WebFetch, TodoWrite, WebSearch
|
||||
model: inherit
|
||||
---
|
||||
|
||||
You are an expert copy editor specializing in Every's house style guide. Your role is to meticulously review text content and suggest edits to ensure compliance with Every's specific editorial standards.
|
||||
|
||||
When reviewing content, you will:
|
||||
|
||||
1. **Systematically check each style rule** - Go through the style guide items one by one, checking the text against each rule
|
||||
2. **Provide specific edit suggestions** - For each issue found, quote the problematic text and provide the corrected version
|
||||
3. **Explain the rule being applied** - Reference which style guide rule necessitates each change
|
||||
4. **Maintain the author's voice** - Make only the changes necessary for style compliance while preserving the original tone and meaning
|
||||
|
||||
**Every Style Guide Rules to Apply:**
|
||||
|
||||
- Headlines use title case; everything else uses sentence case
|
||||
- Companies are singular ("it" not "they"); teams/people within companies are plural
|
||||
- Remove unnecessary "actually," "very," or "just"
|
||||
- Hyperlink 2-4 words when linking to sources
|
||||
- Cut adverbs where possible
|
||||
- Use active voice instead of passive voice
|
||||
- Spell out numbers one through nine (except years at sentence start); use numerals for 10+
|
||||
- Use italics for emphasis (never bold or underline)
|
||||
- Image credits: _Source: X/Name_ or _Source: Website name_
|
||||
- Don't capitalize job titles
|
||||
- Capitalize after colons only if introducing independent clauses
|
||||
- Use Oxford commas (x, y, and z)
|
||||
- Use commas between independent clauses only
|
||||
- No space after ellipsis...
|
||||
- Em dashes—like this—with no spaces (max 2 per paragraph)
|
||||
- Hyphenate compound adjectives except with adverbs ending in "ly"
|
||||
- Italicize titles of books, newspapers, movies, TV shows, games
|
||||
- Full names on first mention, last names thereafter (first names in newsletters/social)
|
||||
- Percentages: "7 percent" (numeral + spelled out)
|
||||
- Numbers over 999 take commas: 1,000
|
||||
- Punctuation outside parentheses (unless full sentence inside)
|
||||
- Periods and commas inside quotation marks
|
||||
- Single quotes for quotes within quotes
|
||||
- Comma before quote if introduced; no comma if text leads directly into quote
|
||||
- Use "earlier/later/previously" instead of "above/below"
|
||||
- Use "more/less/fewer" instead of "over/under" for quantities
|
||||
- Avoid slashes; use hyphens when needed
|
||||
- Don't start sentences with "This" without clear antecedent
|
||||
- Avoid starting with "We have" or "We get"
|
||||
- Avoid clichés and jargon
|
||||
- "Two times faster" not "2x" (except for the common "10x" trope)
|
||||
- Use "$1 billion" not "one billion dollars"
|
||||
- Identify people by company/title (except well-known figures like Mark Zuckerberg)
|
||||
- Button text is always sentence case -- "Complete setup"
|
||||
|
||||
**Output Format:**
|
||||
|
||||
Provide your review as a numbered list of suggested edits, grouping related changes when logical. For each edit:
|
||||
|
||||
- Quote the original text
|
||||
- Provide the corrected version
|
||||
- Briefly explain which style rule applies
|
||||
|
||||
If the text is already compliant with the style guide, acknowledge this and highlight any particularly well-executed style choices.
|
||||
|
||||
Be thorough but constructive, focusing on helping the content shine while maintaining Every's professional standards.
|
||||
@@ -40,7 +40,7 @@ When you receive a comment or review feedback, you will:
|
||||
|
||||
- Maintaining consistency with the existing codebase style and patterns
|
||||
- Ensuring the change doesn't break existing functionality
|
||||
- Following any project-specific guidelines from CLAUDE.md
|
||||
- Following any project-specific guidelines from AGENTS.md (or CLAUDE.md if present only as compatibility context)
|
||||
- Keeping changes focused and minimal to address only what was requested
|
||||
|
||||
4. **Verify the Resolution**: After making changes:
|
||||
|
||||
@@ -1,20 +0,0 @@
|
||||
---
|
||||
name: lfg
|
||||
description: Full autonomous engineering workflow
|
||||
argument-hint: "[feature description]"
|
||||
disable-model-invocation: true
|
||||
---
|
||||
|
||||
Run these slash commands in order. Do not do anything else.
|
||||
|
||||
1. `/ralph-wiggum:ralph-loop "finish all slash commands" --completion-promise "DONE"`
|
||||
2. `/workflows:plan $ARGUMENTS`
|
||||
3. `/compound-engineering:deepen-plan`
|
||||
4. `/workflows:work`
|
||||
5. `/workflows:review`
|
||||
6. `/compound-engineering:resolve_todo_parallel`
|
||||
7. `/compound-engineering:test-browser`
|
||||
8. `/compound-engineering:feature-video`
|
||||
9. Output `<promise>DONE</promise>` when video is in PR
|
||||
|
||||
Start with step 1 now.
|
||||
@@ -1,32 +0,0 @@
|
||||
---
|
||||
name: slfg
|
||||
description: Full autonomous engineering workflow using swarm mode for parallel execution
|
||||
argument-hint: "[feature description]"
|
||||
disable-model-invocation: true
|
||||
---
|
||||
|
||||
Swarm-enabled LFG. Run these steps in order, parallelizing where indicated.
|
||||
|
||||
## Sequential Phase
|
||||
|
||||
1. `/ralph-wiggum:ralph-loop "finish all slash commands" --completion-promise "DONE"`
|
||||
2. `/workflows:plan $ARGUMENTS`
|
||||
3. `/compound-engineering:deepen-plan`
|
||||
4. `/workflows:work` — **Use swarm mode**: Make a Task list and launch an army of agent swarm subagents to build the plan
|
||||
|
||||
## Parallel Phase
|
||||
|
||||
After work completes, launch steps 5 and 6 as **parallel swarm agents** (both only need code to be written):
|
||||
|
||||
5. `/workflows:review` — spawn as background Task agent
|
||||
6. `/compound-engineering:test-browser` — spawn as background Task agent
|
||||
|
||||
Wait for both to complete before continuing.
|
||||
|
||||
## Finalize Phase
|
||||
|
||||
7. `/compound-engineering:resolve_todo_parallel` — resolve any findings from the review
|
||||
8. `/compound-engineering:feature-video` — record the final walkthrough and add to PR
|
||||
9. Output `<promise>DONE</promise>` when video is in PR
|
||||
|
||||
Start with step 1 now.
|
||||
@@ -1,129 +0,0 @@
|
||||
---
|
||||
name: workflows:brainstorm
|
||||
description: Explore requirements and approaches through collaborative dialogue before planning implementation
|
||||
argument-hint: "[feature idea or problem to explore]"
|
||||
---
|
||||
|
||||
# Brainstorm a Feature or Improvement
|
||||
|
||||
**Note: The current year is 2026.** Use this when dating brainstorm documents.
|
||||
|
||||
Brainstorming helps answer **WHAT** to build through collaborative dialogue. It precedes `/workflows:plan`, which answers **HOW** to build it.
|
||||
|
||||
**Process knowledge:** Load the `brainstorming` skill for detailed question techniques, approach exploration patterns, and YAGNI principles.
|
||||
|
||||
## Feature Description
|
||||
|
||||
<feature_description> #$ARGUMENTS </feature_description>
|
||||
|
||||
**If the feature description above is empty, ask the user:** "What would you like to explore? Please describe the feature, problem, or improvement you're thinking about."
|
||||
|
||||
Do not proceed until you have a feature description from the user.
|
||||
|
||||
## Execution Flow
|
||||
|
||||
### Phase 0: Assess Requirements Clarity
|
||||
|
||||
Evaluate whether brainstorming is needed based on the feature description.
|
||||
|
||||
**Clear requirements indicators:**
|
||||
- Specific acceptance criteria provided
|
||||
- Referenced existing patterns to follow
|
||||
- Described exact expected behavior
|
||||
- Constrained, well-defined scope
|
||||
|
||||
**If requirements are already clear:**
|
||||
Use **AskUserQuestion tool** to suggest: "Your requirements seem detailed enough to proceed directly to planning. Should I run `/workflows:plan` instead, or would you like to explore the idea further?"
|
||||
|
||||
### Phase 1: Understand the Idea
|
||||
|
||||
#### 1.1 Repository Research (Lightweight)
|
||||
|
||||
Run a quick repo scan to understand existing patterns:
|
||||
|
||||
- Task repo-research-analyst("Understand existing patterns related to: <feature_description>")
|
||||
|
||||
Focus on: similar features, established patterns, CLAUDE.md guidance.
|
||||
|
||||
#### 1.2 Collaborative Dialogue
|
||||
|
||||
Use the **AskUserQuestion tool** to ask questions **one at a time**.
|
||||
|
||||
**Guidelines (see `brainstorming` skill for detailed techniques):**
|
||||
- Prefer multiple choice when natural options exist
|
||||
- Start broad (purpose, users) then narrow (constraints, edge cases)
|
||||
- Validate assumptions explicitly
|
||||
- Ask about success criteria
|
||||
|
||||
**Exit condition:** Continue until the idea is clear OR user says "proceed"
|
||||
|
||||
### Phase 2: Explore Approaches
|
||||
|
||||
Propose **2-3 concrete approaches** based on research and conversation.
|
||||
|
||||
For each approach, provide:
|
||||
- Brief description (2-3 sentences)
|
||||
- Pros and cons
|
||||
- When it's best suited
|
||||
|
||||
Lead with your recommendation and explain why. Apply YAGNI—prefer simpler solutions.
|
||||
|
||||
Use **AskUserQuestion tool** to ask which approach the user prefers.
|
||||
|
||||
### Phase 3: Capture the Design
|
||||
|
||||
Write a brainstorm document to `docs/brainstorms/YYYY-MM-DD-<topic>-brainstorm.md`.
|
||||
|
||||
**Document structure:** See the `brainstorming` skill for the template format. Key sections: What We're Building, Why This Approach, Key Decisions, Open Questions.
|
||||
|
||||
Ensure `docs/brainstorms/` directory exists before writing.
|
||||
|
||||
**IMPORTANT:** Before proceeding to Phase 4, check if there are any Open Questions listed in the brainstorm document. If there are open questions, YOU MUST ask the user about each one using AskUserQuestion before offering to proceed to planning. Move resolved questions to a "Resolved Questions" section.
|
||||
|
||||
### Phase 4: Handoff
|
||||
|
||||
Use **AskUserQuestion tool** to present next steps:
|
||||
|
||||
**Question:** "Brainstorm captured. What would you like to do next?"
|
||||
|
||||
**Options:**
|
||||
1. **Review and refine** - Improve the document through structured self-review
|
||||
2. **Proceed to planning** - Run `/workflows:plan` (will auto-detect this brainstorm)
|
||||
3. **Ask more questions** - I have more questions to clarify before moving on
|
||||
4. **Done for now** - Return later
|
||||
|
||||
**If user selects "Ask more questions":** YOU (Claude) return to Phase 1.2 (Collaborative Dialogue) and continue asking the USER questions one at a time to further refine the design. The user wants YOU to probe deeper - ask about edge cases, constraints, preferences, or areas not yet explored. Continue until the user is satisfied, then return to Phase 4.
|
||||
|
||||
**If user selects "Review and refine":**
|
||||
|
||||
Load the `document-review` skill and apply it to the brainstorm document.
|
||||
|
||||
When document-review returns "Review complete", present next steps:
|
||||
|
||||
1. **Move to planning** - Continue to `/workflows:plan` with this document
|
||||
2. **Done for now** - Brainstorming complete. To start planning later: `/workflows:plan [document-path]`
|
||||
|
||||
## Output Summary
|
||||
|
||||
When complete, display:
|
||||
|
||||
```
|
||||
Brainstorm complete!
|
||||
|
||||
Document: docs/brainstorms/YYYY-MM-DD-<topic>-brainstorm.md
|
||||
|
||||
Key decisions:
|
||||
- [Decision 1]
|
||||
- [Decision 2]
|
||||
|
||||
Next: Run `/workflows:plan` when ready to implement.
|
||||
```
|
||||
|
||||
## Important Guidelines
|
||||
|
||||
- **Stay focused on WHAT, not HOW** - Implementation details belong in the plan
|
||||
- **Ask one question at a time** - Don't overwhelm
|
||||
- **Apply YAGNI** - Prefer simpler approaches
|
||||
- **Keep outputs concise** - 200-300 words per section max
|
||||
|
||||
NEVER CODE! Just explore and document decisions.
|
||||
@@ -3,9 +3,9 @@ name: agent-browser
|
||||
description: Browser automation using Vercel's agent-browser CLI. Use when you need to interact with web pages, fill forms, take screenshots, or scrape data. Alternative to Playwright MCP - uses Bash commands with ref-based element selection. Triggers on "browse website", "fill form", "click button", "take screenshot", "scrape page", "web automation".
|
||||
---
|
||||
|
||||
# agent-browser: CLI Browser Automation
|
||||
# Browser Automation with agent-browser
|
||||
|
||||
Vercel's headless browser automation CLI designed for AI agents. Uses ref-based selection (@e1, @e2) from accessibility snapshots.
|
||||
The CLI uses Chrome/Chromium via CDP directly. Install via `npm i -g agent-browser`, `brew install agent-browser`, or `cargo install agent-browser`. Run `agent-browser install` to download Chrome.
|
||||
|
||||
## Setup Check
|
||||
|
||||
@@ -23,185 +23,627 @@ agent-browser install # Downloads Chromium
|
||||
|
||||
## Core Workflow
|
||||
|
||||
**The snapshot + ref pattern is optimal for LLMs:**
|
||||
Every browser automation follows this pattern:
|
||||
|
||||
1. **Navigate** to URL
|
||||
2. **Snapshot** to get interactive elements with refs
|
||||
3. **Interact** using refs (@e1, @e2, etc.)
|
||||
4. **Re-snapshot** after navigation or DOM changes
|
||||
1. **Navigate**: `agent-browser open <url>`
|
||||
2. **Snapshot**: `agent-browser snapshot -i` (get element refs like `@e1`, `@e2`)
|
||||
3. **Interact**: Use refs to click, fill, select
|
||||
4. **Re-snapshot**: After navigation or DOM changes, get fresh refs
|
||||
|
||||
```bash
|
||||
# Step 1: Open URL
|
||||
agent-browser open https://example.com
|
||||
|
||||
# Step 2: Get interactive elements with refs
|
||||
agent-browser snapshot -i --json
|
||||
|
||||
# Step 3: Interact using refs
|
||||
agent-browser click @e1
|
||||
agent-browser fill @e2 "search query"
|
||||
|
||||
# Step 4: Re-snapshot after changes
|
||||
agent-browser open https://example.com/form
|
||||
agent-browser snapshot -i
|
||||
```
|
||||
# Output: @e1 [input type="email"], @e2 [input type="password"], @e3 [button] "Submit"
|
||||
|
||||
## Key Commands
|
||||
|
||||
### Navigation
|
||||
|
||||
```bash
|
||||
agent-browser open <url> # Navigate to URL
|
||||
agent-browser back # Go back
|
||||
agent-browser forward # Go forward
|
||||
agent-browser reload # Reload page
|
||||
agent-browser close # Close browser
|
||||
```
|
||||
|
||||
### Snapshots (Essential for AI)
|
||||
|
||||
```bash
|
||||
agent-browser snapshot # Full accessibility tree
|
||||
agent-browser snapshot -i # Interactive elements only (recommended)
|
||||
agent-browser snapshot -i --json # JSON output for parsing
|
||||
agent-browser snapshot -c # Compact (remove empty elements)
|
||||
agent-browser snapshot -d 3 # Limit depth
|
||||
```
|
||||
|
||||
### Interactions
|
||||
|
||||
```bash
|
||||
agent-browser click @e1 # Click element
|
||||
agent-browser dblclick @e1 # Double-click
|
||||
agent-browser fill @e1 "text" # Clear and fill input
|
||||
agent-browser type @e1 "text" # Type without clearing
|
||||
agent-browser press Enter # Press key
|
||||
agent-browser hover @e1 # Hover element
|
||||
agent-browser check @e1 # Check checkbox
|
||||
agent-browser uncheck @e1 # Uncheck checkbox
|
||||
agent-browser select @e1 "option" # Select dropdown option
|
||||
agent-browser scroll down 500 # Scroll (up/down/left/right)
|
||||
agent-browser scrollintoview @e1 # Scroll element into view
|
||||
```
|
||||
|
||||
### Get Information
|
||||
|
||||
```bash
|
||||
agent-browser get text @e1 # Get element text
|
||||
agent-browser get html @e1 # Get element HTML
|
||||
agent-browser get value @e1 # Get input value
|
||||
agent-browser get attr href @e1 # Get attribute
|
||||
agent-browser get title # Get page title
|
||||
agent-browser get url # Get current URL
|
||||
agent-browser get count "button" # Count matching elements
|
||||
```
|
||||
|
||||
### Screenshots & PDFs
|
||||
|
||||
```bash
|
||||
agent-browser screenshot # Viewport screenshot
|
||||
agent-browser screenshot --full # Full page
|
||||
agent-browser screenshot output.png # Save to file
|
||||
agent-browser screenshot --full output.png # Full page to file
|
||||
agent-browser pdf output.pdf # Save as PDF
|
||||
```
|
||||
|
||||
### Wait
|
||||
|
||||
```bash
|
||||
agent-browser wait @e1 # Wait for element
|
||||
agent-browser wait 2000 # Wait milliseconds
|
||||
agent-browser wait "text" # Wait for text to appear
|
||||
```
|
||||
|
||||
## Semantic Locators (Alternative to Refs)
|
||||
|
||||
```bash
|
||||
agent-browser find role button click --name "Submit"
|
||||
agent-browser find text "Sign up" click
|
||||
agent-browser find label "Email" fill "user@example.com"
|
||||
agent-browser find placeholder "Search..." fill "query"
|
||||
```
|
||||
|
||||
## Sessions (Parallel Browsers)
|
||||
|
||||
```bash
|
||||
# Run multiple independent browser sessions
|
||||
agent-browser --session browser1 open https://site1.com
|
||||
agent-browser --session browser2 open https://site2.com
|
||||
|
||||
# List active sessions
|
||||
agent-browser session list
|
||||
```
|
||||
|
||||
## Examples
|
||||
|
||||
### Login Flow
|
||||
|
||||
```bash
|
||||
agent-browser open https://app.example.com/login
|
||||
agent-browser snapshot -i
|
||||
# Output shows: textbox "Email" [ref=e1], textbox "Password" [ref=e2], button "Sign in" [ref=e3]
|
||||
agent-browser fill @e1 "user@example.com"
|
||||
agent-browser fill @e2 "password123"
|
||||
agent-browser click @e3
|
||||
agent-browser wait 2000
|
||||
agent-browser snapshot -i # Verify logged in
|
||||
agent-browser wait --load networkidle
|
||||
agent-browser snapshot -i # Check result
|
||||
```
|
||||
|
||||
### Search and Extract
|
||||
## Command Chaining
|
||||
|
||||
Commands can be chained with `&&` in a single shell invocation. The browser persists between commands via a background daemon, so chaining is safe and more efficient than separate calls.
|
||||
|
||||
```bash
|
||||
agent-browser open https://news.ycombinator.com
|
||||
agent-browser snapshot -i --json
|
||||
# Parse JSON to find story links
|
||||
agent-browser get text @e12 # Get headline text
|
||||
agent-browser click @e12 # Click to open story
|
||||
# Chain open + wait + snapshot in one call
|
||||
agent-browser open https://example.com && agent-browser wait --load networkidle && agent-browser snapshot -i
|
||||
|
||||
# Chain multiple interactions
|
||||
agent-browser fill @e1 "user@example.com" && agent-browser fill @e2 "password123" && agent-browser click @e3
|
||||
|
||||
# Navigate and capture
|
||||
agent-browser open https://example.com && agent-browser wait --load networkidle && agent-browser screenshot page.png
|
||||
```
|
||||
|
||||
### Form Filling
|
||||
**When to chain:** Use `&&` when you don't need to read the output of an intermediate command before proceeding (e.g., open + wait + screenshot). Run commands separately when you need to parse the output first (e.g., snapshot to discover refs, then interact using those refs).
|
||||
|
||||
## Handling Authentication
|
||||
|
||||
When automating a site that requires login, choose the approach that fits:
|
||||
|
||||
**Option 1: Import auth from the user's browser (fastest for one-off tasks)**
|
||||
|
||||
```bash
|
||||
agent-browser open https://forms.example.com
|
||||
# Connect to the user's running Chrome (they're already logged in)
|
||||
agent-browser --auto-connect state save ./auth.json
|
||||
# Use that auth state
|
||||
agent-browser --state ./auth.json open https://app.example.com/dashboard
|
||||
```
|
||||
|
||||
State files contain session tokens in plaintext -- add to `.gitignore` and delete when no longer needed. Set `AGENT_BROWSER_ENCRYPTION_KEY` for encryption at rest.
|
||||
|
||||
**Option 2: Persistent profile (simplest for recurring tasks)**
|
||||
|
||||
```bash
|
||||
# First run: login manually or via automation
|
||||
agent-browser --profile ~/.myapp open https://app.example.com/login
|
||||
# ... fill credentials, submit ...
|
||||
|
||||
# All future runs: already authenticated
|
||||
agent-browser --profile ~/.myapp open https://app.example.com/dashboard
|
||||
```
|
||||
|
||||
**Option 3: Session name (auto-save/restore cookies + localStorage)**
|
||||
|
||||
```bash
|
||||
agent-browser --session-name myapp open https://app.example.com/login
|
||||
# ... login flow ...
|
||||
agent-browser close # State auto-saved
|
||||
|
||||
# Next time: state auto-restored
|
||||
agent-browser --session-name myapp open https://app.example.com/dashboard
|
||||
```
|
||||
|
||||
**Option 4: Auth vault (credentials stored encrypted, login by name)**
|
||||
|
||||
```bash
|
||||
echo "$PASSWORD" | agent-browser auth save myapp --url https://app.example.com/login --username user --password-stdin
|
||||
agent-browser auth login myapp
|
||||
```
|
||||
|
||||
**Option 5: State file (manual save/load)**
|
||||
|
||||
```bash
|
||||
# After logging in:
|
||||
agent-browser state save ./auth.json
|
||||
# In a future session:
|
||||
agent-browser state load ./auth.json
|
||||
agent-browser open https://app.example.com/dashboard
|
||||
```
|
||||
|
||||
See [references/authentication.md](references/authentication.md) for OAuth, 2FA, cookie-based auth, and token refresh patterns.
|
||||
|
||||
## Essential Commands
|
||||
|
||||
```bash
|
||||
# Navigation
|
||||
agent-browser open <url> # Navigate (aliases: goto, navigate)
|
||||
agent-browser close # Close browser
|
||||
|
||||
# Snapshot
|
||||
agent-browser snapshot -i # Interactive elements with refs (recommended)
|
||||
agent-browser snapshot -i -C # Include cursor-interactive elements (divs with onclick, cursor:pointer)
|
||||
agent-browser snapshot -s "#selector" # Scope to CSS selector
|
||||
|
||||
# Interaction (use @refs from snapshot)
|
||||
agent-browser click @e1 # Click element
|
||||
agent-browser click @e1 --new-tab # Click and open in new tab
|
||||
agent-browser fill @e2 "text" # Clear and type text
|
||||
agent-browser type @e2 "text" # Type without clearing
|
||||
agent-browser select @e1 "option" # Select dropdown option
|
||||
agent-browser check @e1 # Check checkbox
|
||||
agent-browser press Enter # Press key
|
||||
agent-browser keyboard type "text" # Type at current focus (no selector)
|
||||
agent-browser keyboard inserttext "text" # Insert without key events
|
||||
agent-browser scroll down 500 # Scroll page
|
||||
agent-browser scroll down 500 --selector "div.content" # Scroll within a specific container
|
||||
|
||||
# Get information
|
||||
agent-browser get text @e1 # Get element text
|
||||
agent-browser get url # Get current URL
|
||||
agent-browser get title # Get page title
|
||||
agent-browser get cdp-url # Get CDP WebSocket URL
|
||||
|
||||
# Wait
|
||||
agent-browser wait @e1 # Wait for element
|
||||
agent-browser wait --load networkidle # Wait for network idle
|
||||
agent-browser wait --url "**/page" # Wait for URL pattern
|
||||
agent-browser wait 2000 # Wait milliseconds
|
||||
agent-browser wait --text "Welcome" # Wait for text to appear (substring match)
|
||||
agent-browser wait --fn "!document.body.innerText.includes('Loading...')" # Wait for text to disappear
|
||||
agent-browser wait "#spinner" --state hidden # Wait for element to disappear
|
||||
|
||||
# Downloads
|
||||
agent-browser download @e1 ./file.pdf # Click element to trigger download
|
||||
agent-browser wait --download ./output.zip # Wait for any download to complete
|
||||
agent-browser --download-path ./downloads open <url> # Set default download directory
|
||||
|
||||
# Viewport & Device Emulation
|
||||
agent-browser set viewport 1920 1080 # Set viewport size (default: 1280x720)
|
||||
agent-browser set viewport 1920 1080 2 # 2x retina (same CSS size, higher res screenshots)
|
||||
agent-browser set device "iPhone 14" # Emulate device (viewport + user agent)
|
||||
|
||||
# Capture
|
||||
agent-browser screenshot # Screenshot to temp dir
|
||||
agent-browser screenshot --full # Full page screenshot
|
||||
agent-browser screenshot --annotate # Annotated screenshot with numbered element labels
|
||||
agent-browser screenshot --screenshot-dir ./shots # Save to custom directory
|
||||
agent-browser screenshot --screenshot-format jpeg --screenshot-quality 80
|
||||
agent-browser pdf output.pdf # Save as PDF
|
||||
|
||||
# Clipboard
|
||||
agent-browser clipboard read # Read text from clipboard
|
||||
agent-browser clipboard write "Hello, World!" # Write text to clipboard
|
||||
agent-browser clipboard copy # Copy current selection
|
||||
agent-browser clipboard paste # Paste from clipboard
|
||||
|
||||
# Diff (compare page states)
|
||||
agent-browser diff snapshot # Compare current vs last snapshot
|
||||
agent-browser diff snapshot --baseline before.txt # Compare current vs saved file
|
||||
agent-browser diff screenshot --baseline before.png # Visual pixel diff
|
||||
agent-browser diff url <url1> <url2> # Compare two pages
|
||||
agent-browser diff url <url1> <url2> --wait-until networkidle # Custom wait strategy
|
||||
agent-browser diff url <url1> <url2> --selector "#main" # Scope to element
|
||||
```
|
||||
|
||||
## Common Patterns
|
||||
|
||||
### Form Submission
|
||||
|
||||
```bash
|
||||
agent-browser open https://example.com/signup
|
||||
agent-browser snapshot -i
|
||||
agent-browser fill @e1 "John Doe"
|
||||
agent-browser fill @e2 "john@example.com"
|
||||
agent-browser select @e3 "United States"
|
||||
agent-browser check @e4 # Agree to terms
|
||||
agent-browser click @e5 # Submit button
|
||||
agent-browser screenshot confirmation.png
|
||||
agent-browser fill @e1 "Jane Doe"
|
||||
agent-browser fill @e2 "jane@example.com"
|
||||
agent-browser select @e3 "California"
|
||||
agent-browser check @e4
|
||||
agent-browser click @e5
|
||||
agent-browser wait --load networkidle
|
||||
```
|
||||
|
||||
### Debug Mode
|
||||
### Authentication with Auth Vault (Recommended)
|
||||
|
||||
```bash
|
||||
# Run with visible browser window
|
||||
agent-browser --headed open https://example.com
|
||||
agent-browser --headed snapshot -i
|
||||
agent-browser --headed click @e1
|
||||
# Save credentials once (encrypted with AGENT_BROWSER_ENCRYPTION_KEY)
|
||||
# Recommended: pipe password via stdin to avoid shell history exposure
|
||||
echo "pass" | agent-browser auth save github --url https://github.com/login --username user --password-stdin
|
||||
|
||||
# Login using saved profile (LLM never sees password)
|
||||
agent-browser auth login github
|
||||
|
||||
# List/show/delete profiles
|
||||
agent-browser auth list
|
||||
agent-browser auth show github
|
||||
agent-browser auth delete github
|
||||
```
|
||||
|
||||
## JSON Output
|
||||
|
||||
Add `--json` for structured output:
|
||||
### Authentication with State Persistence
|
||||
|
||||
```bash
|
||||
# Login once and save state
|
||||
agent-browser open https://app.example.com/login
|
||||
agent-browser snapshot -i
|
||||
agent-browser fill @e1 "$USERNAME"
|
||||
agent-browser fill @e2 "$PASSWORD"
|
||||
agent-browser click @e3
|
||||
agent-browser wait --url "**/dashboard"
|
||||
agent-browser state save auth.json
|
||||
|
||||
# Reuse in future sessions
|
||||
agent-browser state load auth.json
|
||||
agent-browser open https://app.example.com/dashboard
|
||||
```
|
||||
|
||||
### Session Persistence
|
||||
|
||||
```bash
|
||||
# Auto-save/restore cookies and localStorage across browser restarts
|
||||
agent-browser --session-name myapp open https://app.example.com/login
|
||||
# ... login flow ...
|
||||
agent-browser close # State auto-saved to ~/.agent-browser/sessions/
|
||||
|
||||
# Next time, state is auto-loaded
|
||||
agent-browser --session-name myapp open https://app.example.com/dashboard
|
||||
|
||||
# Encrypt state at rest
|
||||
export AGENT_BROWSER_ENCRYPTION_KEY=$(openssl rand -hex 32)
|
||||
agent-browser --session-name secure open https://app.example.com
|
||||
|
||||
# Manage saved states
|
||||
agent-browser state list
|
||||
agent-browser state show myapp-default.json
|
||||
agent-browser state clear myapp
|
||||
agent-browser state clean --older-than 7
|
||||
```
|
||||
|
||||
### Data Extraction
|
||||
|
||||
```bash
|
||||
agent-browser open https://example.com/products
|
||||
agent-browser snapshot -i
|
||||
agent-browser get text @e5 # Get specific element text
|
||||
agent-browser get text body > page.txt # Get all page text
|
||||
|
||||
# JSON output for parsing
|
||||
agent-browser snapshot -i --json
|
||||
agent-browser get text @e1 --json
|
||||
```
|
||||
|
||||
Returns:
|
||||
### Parallel Sessions
|
||||
|
||||
```bash
|
||||
agent-browser --session site1 open https://site-a.com
|
||||
agent-browser --session site2 open https://site-b.com
|
||||
|
||||
agent-browser --session site1 snapshot -i
|
||||
agent-browser --session site2 snapshot -i
|
||||
|
||||
agent-browser session list
|
||||
```
|
||||
|
||||
### Connect to Existing Chrome
|
||||
|
||||
```bash
|
||||
# Auto-discover running Chrome with remote debugging enabled
|
||||
agent-browser --auto-connect open https://example.com
|
||||
agent-browser --auto-connect snapshot
|
||||
|
||||
# Or with explicit CDP port
|
||||
agent-browser --cdp 9222 snapshot
|
||||
```
|
||||
|
||||
### Color Scheme (Dark Mode)
|
||||
|
||||
```bash
|
||||
# Persistent dark mode via flag (applies to all pages and new tabs)
|
||||
agent-browser --color-scheme dark open https://example.com
|
||||
|
||||
# Or via environment variable
|
||||
AGENT_BROWSER_COLOR_SCHEME=dark agent-browser open https://example.com
|
||||
|
||||
# Or set during session (persists for subsequent commands)
|
||||
agent-browser set media dark
|
||||
```
|
||||
|
||||
### Viewport & Responsive Testing
|
||||
|
||||
```bash
|
||||
# Set a custom viewport size (default is 1280x720)
|
||||
agent-browser set viewport 1920 1080
|
||||
agent-browser screenshot desktop.png
|
||||
|
||||
# Test mobile-width layout
|
||||
agent-browser set viewport 375 812
|
||||
agent-browser screenshot mobile.png
|
||||
|
||||
# Retina/HiDPI: same CSS layout at 2x pixel density
|
||||
# Screenshots stay at logical viewport size, but content renders at higher DPI
|
||||
agent-browser set viewport 1920 1080 2
|
||||
agent-browser screenshot retina.png
|
||||
|
||||
# Device emulation (sets viewport + user agent in one step)
|
||||
agent-browser set device "iPhone 14"
|
||||
agent-browser screenshot device.png
|
||||
```
|
||||
|
||||
The `scale` parameter (3rd argument) sets `window.devicePixelRatio` without changing CSS layout. Use it when testing retina rendering or capturing higher-resolution screenshots.
|
||||
|
||||
### Visual Browser (Debugging)
|
||||
|
||||
```bash
|
||||
agent-browser --headed open https://example.com
|
||||
agent-browser highlight @e1 # Highlight element
|
||||
agent-browser inspect # Open Chrome DevTools for the active page
|
||||
agent-browser record start demo.webm # Record session
|
||||
agent-browser profiler start # Start Chrome DevTools profiling
|
||||
agent-browser profiler stop trace.json # Stop and save profile (path optional)
|
||||
```
|
||||
|
||||
Use `AGENT_BROWSER_HEADED=1` to enable headed mode via environment variable. Browser extensions work in both headed and headless mode.
|
||||
|
||||
### Local Files (PDFs, HTML)
|
||||
|
||||
```bash
|
||||
# Open local files with file:// URLs
|
||||
agent-browser --allow-file-access open file:///path/to/document.pdf
|
||||
agent-browser --allow-file-access open file:///path/to/page.html
|
||||
agent-browser screenshot output.png
|
||||
```
|
||||
|
||||
### iOS Simulator (Mobile Safari)
|
||||
|
||||
```bash
|
||||
# List available iOS simulators
|
||||
agent-browser device list
|
||||
|
||||
# Launch Safari on a specific device
|
||||
agent-browser -p ios --device "iPhone 16 Pro" open https://example.com
|
||||
|
||||
# Same workflow as desktop - snapshot, interact, re-snapshot
|
||||
agent-browser -p ios snapshot -i
|
||||
agent-browser -p ios tap @e1 # Tap (alias for click)
|
||||
agent-browser -p ios fill @e2 "text"
|
||||
agent-browser -p ios swipe up # Mobile-specific gesture
|
||||
|
||||
# Take screenshot
|
||||
agent-browser -p ios screenshot mobile.png
|
||||
|
||||
# Close session (shuts down simulator)
|
||||
agent-browser -p ios close
|
||||
```
|
||||
|
||||
**Requirements:** macOS with Xcode, Appium (`npm install -g appium && appium driver install xcuitest`)
|
||||
|
||||
**Real devices:** Works with physical iOS devices if pre-configured. Use `--device "<UDID>"` where UDID is from `xcrun xctrace list devices`.
|
||||
|
||||
## Security
|
||||
|
||||
All security features are opt-in. By default, agent-browser imposes no restrictions on navigation, actions, or output.
|
||||
|
||||
### Content Boundaries (Recommended for AI Agents)
|
||||
|
||||
Enable `--content-boundaries` to wrap page-sourced output in markers that help LLMs distinguish tool output from untrusted page content:
|
||||
|
||||
```bash
|
||||
export AGENT_BROWSER_CONTENT_BOUNDARIES=1
|
||||
agent-browser snapshot
|
||||
# Output:
|
||||
# --- AGENT_BROWSER_PAGE_CONTENT nonce=<hex> origin=https://example.com ---
|
||||
# [accessibility tree]
|
||||
# --- END_AGENT_BROWSER_PAGE_CONTENT nonce=<hex> ---
|
||||
```
|
||||
|
||||
### Domain Allowlist
|
||||
|
||||
Restrict navigation to trusted domains. Wildcards like `*.example.com` also match the bare domain `example.com`. Sub-resource requests, WebSocket, and EventSource connections to non-allowed domains are also blocked. Include CDN domains your target pages depend on:
|
||||
|
||||
```bash
|
||||
export AGENT_BROWSER_ALLOWED_DOMAINS="example.com,*.example.com"
|
||||
agent-browser open https://example.com # OK
|
||||
agent-browser open https://malicious.com # Blocked
|
||||
```
|
||||
|
||||
### Action Policy
|
||||
|
||||
Use a policy file to gate destructive actions:
|
||||
|
||||
```bash
|
||||
export AGENT_BROWSER_ACTION_POLICY=./policy.json
|
||||
```
|
||||
|
||||
Example `policy.json`:
|
||||
|
||||
```json
|
||||
{ "default": "deny", "allow": ["navigate", "snapshot", "click", "scroll", "wait", "get"] }
|
||||
```
|
||||
|
||||
Auth vault operations (`auth login`, etc.) bypass action policy but domain allowlist still applies.
|
||||
|
||||
### Output Limits
|
||||
|
||||
Prevent context flooding from large pages:
|
||||
|
||||
```bash
|
||||
export AGENT_BROWSER_MAX_OUTPUT=50000
|
||||
```
|
||||
|
||||
## Diffing (Verifying Changes)
|
||||
|
||||
Use `diff snapshot` after performing an action to verify it had the intended effect. This compares the current accessibility tree against the last snapshot taken in the session.
|
||||
|
||||
```bash
|
||||
# Typical workflow: snapshot -> action -> diff
|
||||
agent-browser snapshot -i # Take baseline snapshot
|
||||
agent-browser click @e2 # Perform action
|
||||
agent-browser diff snapshot # See what changed (auto-compares to last snapshot)
|
||||
```
|
||||
|
||||
For visual regression testing or monitoring:
|
||||
|
||||
```bash
|
||||
# Save a baseline screenshot, then compare later
|
||||
agent-browser screenshot baseline.png
|
||||
# ... time passes or changes are made ...
|
||||
agent-browser diff screenshot --baseline baseline.png
|
||||
|
||||
# Compare staging vs production
|
||||
agent-browser diff url https://staging.example.com https://prod.example.com --screenshot
|
||||
```
|
||||
|
||||
`diff snapshot` output uses `+` for additions and `-` for removals, similar to git diff. `diff screenshot` produces a diff image with changed pixels highlighted in red, plus a mismatch percentage.
|
||||
|
||||
## Timeouts and Slow Pages
|
||||
|
||||
The default timeout is 25 seconds. This can be overridden with the `AGENT_BROWSER_DEFAULT_TIMEOUT` environment variable (value in milliseconds). For slow websites or large pages, use explicit waits instead of relying on the default timeout:
|
||||
|
||||
```bash
|
||||
# Wait for network activity to settle (best for slow pages)
|
||||
agent-browser wait --load networkidle
|
||||
|
||||
# Wait for a specific element to appear
|
||||
agent-browser wait "#content"
|
||||
agent-browser wait @e1
|
||||
|
||||
# Wait for a specific URL pattern (useful after redirects)
|
||||
agent-browser wait --url "**/dashboard"
|
||||
|
||||
# Wait for a JavaScript condition
|
||||
agent-browser wait --fn "document.readyState === 'complete'"
|
||||
|
||||
# Wait a fixed duration (milliseconds) as a last resort
|
||||
agent-browser wait 5000
|
||||
```
|
||||
|
||||
When dealing with consistently slow websites, use `wait --load networkidle` after `open` to ensure the page is fully loaded before taking a snapshot. If a specific element is slow to render, wait for it directly with `wait <selector>` or `wait @ref`.
|
||||
|
||||
## Session Management and Cleanup
|
||||
|
||||
When running multiple agents or automations concurrently, always use named sessions to avoid conflicts:
|
||||
|
||||
```bash
|
||||
# Each agent gets its own isolated session
|
||||
agent-browser --session agent1 open site-a.com
|
||||
agent-browser --session agent2 open site-b.com
|
||||
|
||||
# Check active sessions
|
||||
agent-browser session list
|
||||
```
|
||||
|
||||
Always close your browser session when done to avoid leaked processes:
|
||||
|
||||
```bash
|
||||
agent-browser close # Close default session
|
||||
agent-browser --session agent1 close # Close specific session
|
||||
```
|
||||
|
||||
If a previous session was not closed properly, the daemon may still be running. Use `agent-browser close` to clean it up before starting new work.
|
||||
|
||||
To auto-shutdown the daemon after a period of inactivity (useful for ephemeral/CI environments):
|
||||
|
||||
```bash
|
||||
AGENT_BROWSER_IDLE_TIMEOUT_MS=60000 agent-browser open example.com
|
||||
```
|
||||
|
||||
## Ref Lifecycle (Important)
|
||||
|
||||
Refs (`@e1`, `@e2`, etc.) are invalidated when the page changes. Always re-snapshot after:
|
||||
|
||||
- Clicking links or buttons that navigate
|
||||
- Form submissions
|
||||
- Dynamic content loading (dropdowns, modals)
|
||||
|
||||
```bash
|
||||
agent-browser click @e5 # Navigates to new page
|
||||
agent-browser snapshot -i # MUST re-snapshot
|
||||
agent-browser click @e1 # Use new refs
|
||||
```
|
||||
|
||||
## Annotated Screenshots (Vision Mode)
|
||||
|
||||
Use `--annotate` to take a screenshot with numbered labels overlaid on interactive elements. Each label `[N]` maps to ref `@eN`. This also caches refs, so you can interact with elements immediately without a separate snapshot.
|
||||
|
||||
```bash
|
||||
agent-browser screenshot --annotate
|
||||
# Output includes the image path and a legend:
|
||||
# [1] @e1 button "Submit"
|
||||
# [2] @e2 link "Home"
|
||||
# [3] @e3 textbox "Email"
|
||||
agent-browser click @e2 # Click using ref from annotated screenshot
|
||||
```
|
||||
|
||||
Use annotated screenshots when:
|
||||
|
||||
- The page has unlabeled icon buttons or visual-only elements
|
||||
- You need to verify visual layout or styling
|
||||
- Canvas or chart elements are present (invisible to text snapshots)
|
||||
- You need spatial reasoning about element positions
|
||||
|
||||
## Semantic Locators (Alternative to Refs)
|
||||
|
||||
When refs are unavailable or unreliable, use semantic locators:
|
||||
|
||||
```bash
|
||||
agent-browser find text "Sign In" click
|
||||
agent-browser find label "Email" fill "user@test.com"
|
||||
agent-browser find role button click --name "Submit"
|
||||
agent-browser find placeholder "Search" type "query"
|
||||
agent-browser find testid "submit-btn" click
|
||||
```
|
||||
|
||||
## JavaScript Evaluation (eval)
|
||||
|
||||
Use `eval` to run JavaScript in the browser context. **Shell quoting can corrupt complex expressions** -- use `--stdin` or `-b` to avoid issues.
|
||||
|
||||
```bash
|
||||
# Simple expressions work with regular quoting
|
||||
agent-browser eval 'document.title'
|
||||
agent-browser eval 'document.querySelectorAll("img").length'
|
||||
|
||||
# Complex JS: use --stdin with heredoc (RECOMMENDED)
|
||||
agent-browser eval --stdin <<'EVALEOF'
|
||||
JSON.stringify(
|
||||
Array.from(document.querySelectorAll("img"))
|
||||
.filter(i => !i.alt)
|
||||
.map(i => ({ src: i.src.split("/").pop(), width: i.width }))
|
||||
)
|
||||
EVALEOF
|
||||
|
||||
# Alternative: base64 encoding (avoids all shell escaping issues)
|
||||
agent-browser eval -b "$(echo -n 'Array.from(document.querySelectorAll("a")).map(a => a.href)' | base64)"
|
||||
```
|
||||
|
||||
**Why this matters:** When the shell processes your command, inner double quotes, `!` characters (history expansion), backticks, and `$()` can all corrupt the JavaScript before it reaches agent-browser. The `--stdin` and `-b` flags bypass shell interpretation entirely.
|
||||
|
||||
**Rules of thumb:**
|
||||
|
||||
- Single-line, no nested quotes -> regular `eval 'expression'` with single quotes is fine
|
||||
- Nested quotes, arrow functions, template literals, or multiline -> use `eval --stdin <<'EVALEOF'`
|
||||
- Programmatic/generated scripts -> use `eval -b` with base64
|
||||
|
||||
## Configuration File
|
||||
|
||||
Create `agent-browser.json` in the project root for persistent settings:
|
||||
|
||||
```json
|
||||
{
|
||||
"success": true,
|
||||
"data": {
|
||||
"refs": {
|
||||
"e1": {"name": "Submit", "role": "button"},
|
||||
"e2": {"name": "Email", "role": "textbox"}
|
||||
},
|
||||
"snapshot": "- button \"Submit\" [ref=e1]\n- textbox \"Email\" [ref=e2]"
|
||||
}
|
||||
"headed": true,
|
||||
"proxy": "http://localhost:8080",
|
||||
"profile": "./browser-data"
|
||||
}
|
||||
```
|
||||
|
||||
Priority (lowest to highest): `~/.agent-browser/config.json` < `./agent-browser.json` < env vars < CLI flags. Use `--config <path>` or `AGENT_BROWSER_CONFIG` env var for a custom config file (exits with error if missing/invalid). All CLI options map to camelCase keys (e.g., `--executable-path` -> `"executablePath"`). Boolean flags accept `true`/`false` values (e.g., `--headed false` overrides config). Extensions from user and project configs are merged, not replaced.
|
||||
|
||||
## Browser Engine Selection
|
||||
|
||||
Use `--engine` to choose a local browser engine. The default is `chrome`.
|
||||
|
||||
```bash
|
||||
# Use Lightpanda (fast headless browser, requires separate install)
|
||||
agent-browser --engine lightpanda open example.com
|
||||
|
||||
# Via environment variable
|
||||
export AGENT_BROWSER_ENGINE=lightpanda
|
||||
agent-browser open example.com
|
||||
|
||||
# With custom binary path
|
||||
agent-browser --engine lightpanda --executable-path /path/to/lightpanda open example.com
|
||||
```
|
||||
|
||||
Supported engines:
|
||||
- `chrome` (default) -- Chrome/Chromium via CDP
|
||||
- `lightpanda` -- Lightpanda headless browser via CDP (10x faster, 10x less memory than Chrome)
|
||||
|
||||
Lightpanda does not support `--extension`, `--profile`, `--state`, or `--allow-file-access`. Install Lightpanda from https://lightpanda.io/docs/open-source/installation.
|
||||
|
||||
## Deep-Dive Documentation
|
||||
|
||||
| Reference | When to Use |
|
||||
| -------------------------------------------------------------------- | --------------------------------------------------------- |
|
||||
| [references/commands.md](references/commands.md) | Full command reference with all options |
|
||||
| [references/snapshot-refs.md](references/snapshot-refs.md) | Ref lifecycle, invalidation rules, troubleshooting |
|
||||
| [references/session-management.md](references/session-management.md) | Parallel sessions, state persistence, concurrent scraping |
|
||||
| [references/authentication.md](references/authentication.md) | Login flows, OAuth, 2FA handling, state reuse |
|
||||
| [references/video-recording.md](references/video-recording.md) | Recording workflows for debugging and documentation |
|
||||
| [references/profiling.md](references/profiling.md) | Chrome DevTools profiling for performance analysis |
|
||||
| [references/proxy-support.md](references/proxy-support.md) | Proxy configuration, geo-testing, rotating proxies |
|
||||
|
||||
## Ready-to-Use Templates
|
||||
|
||||
| Template | Description |
|
||||
| ------------------------------------------------------------------------ | ----------------------------------- |
|
||||
| [templates/form-automation.sh](templates/form-automation.sh) | Form filling with validation |
|
||||
| [templates/authenticated-session.sh](templates/authenticated-session.sh) | Login once, reuse state |
|
||||
| [templates/capture-workflow.sh](templates/capture-workflow.sh) | Content extraction with screenshots |
|
||||
|
||||
```bash
|
||||
./templates/form-automation.sh https://example.com/form
|
||||
./templates/authenticated-session.sh https://app.example.com/login
|
||||
./templates/capture-workflow.sh https://example.com ./output
|
||||
```
|
||||
|
||||
## vs Playwright MCP
|
||||
|
||||
| Feature | agent-browser (CLI) | Playwright MCP |
|
||||
|
||||
@@ -0,0 +1,303 @@
|
||||
# Authentication Patterns
|
||||
|
||||
Login flows, session persistence, OAuth, 2FA, and authenticated browsing.
|
||||
|
||||
**Related**: [commands.md](commands.md) for full command reference, [SKILL.md](../SKILL.md) for quick start.
|
||||
|
||||
## Contents
|
||||
|
||||
- [Import Auth from Your Browser](#import-auth-from-your-browser)
|
||||
- [Persistent Profiles](#persistent-profiles)
|
||||
- [Session Persistence](#session-persistence)
|
||||
- [Basic Login Flow](#basic-login-flow)
|
||||
- [Saving Authentication State](#saving-authentication-state)
|
||||
- [Restoring Authentication](#restoring-authentication)
|
||||
- [OAuth / SSO Flows](#oauth--sso-flows)
|
||||
- [Two-Factor Authentication](#two-factor-authentication)
|
||||
- [HTTP Basic Auth](#http-basic-auth)
|
||||
- [Cookie-Based Auth](#cookie-based-auth)
|
||||
- [Token Refresh Handling](#token-refresh-handling)
|
||||
- [Security Best Practices](#security-best-practices)
|
||||
|
||||
## Import Auth from Your Browser
|
||||
|
||||
The fastest way to authenticate is to reuse cookies from a Chrome session you are already logged into.
|
||||
|
||||
**Step 1: Start Chrome with remote debugging**
|
||||
|
||||
```bash
|
||||
# macOS
|
||||
"/Applications/Google Chrome.app/Contents/MacOS/Google Chrome" --remote-debugging-port=9222
|
||||
|
||||
# Linux
|
||||
google-chrome --remote-debugging-port=9222
|
||||
|
||||
# Windows
|
||||
"C:\Program Files\Google\Chrome\Application\chrome.exe" --remote-debugging-port=9222
|
||||
```
|
||||
|
||||
Log in to your target site(s) in this Chrome window as you normally would.
|
||||
|
||||
> **Security note:** `--remote-debugging-port` exposes full browser control on localhost. Any local process can connect and read cookies, execute JS, etc. Only use on trusted machines and close Chrome when done.
|
||||
|
||||
**Step 2: Grab the auth state**
|
||||
|
||||
```bash
|
||||
# Auto-discover the running Chrome and save its cookies + localStorage
|
||||
agent-browser --auto-connect state save ./my-auth.json
|
||||
```
|
||||
|
||||
**Step 3: Reuse in automation**
|
||||
|
||||
```bash
|
||||
# Load auth at launch
|
||||
agent-browser --state ./my-auth.json open https://app.example.com/dashboard
|
||||
|
||||
# Or load into an existing session
|
||||
agent-browser state load ./my-auth.json
|
||||
agent-browser open https://app.example.com/dashboard
|
||||
```
|
||||
|
||||
This works for any site, including those with complex OAuth flows, SSO, or 2FA -- as long as Chrome already has valid session cookies.
|
||||
|
||||
> **Security note:** State files contain session tokens in plaintext. Add them to `.gitignore`, delete when no longer needed, and set `AGENT_BROWSER_ENCRYPTION_KEY` for encryption at rest. See [Security Best Practices](#security-best-practices).
|
||||
|
||||
**Tip:** Combine with `--session-name` so the imported auth auto-persists across restarts:
|
||||
|
||||
```bash
|
||||
agent-browser --session-name myapp state load ./my-auth.json
|
||||
# From now on, state is auto-saved/restored for "myapp"
|
||||
```
|
||||
|
||||
## Persistent Profiles
|
||||
|
||||
Use `--profile` to point agent-browser at a Chrome user data directory. This persists everything (cookies, IndexedDB, service workers, cache) across browser restarts without explicit save/load:
|
||||
|
||||
```bash
|
||||
# First run: login once
|
||||
agent-browser --profile ~/.myapp-profile open https://app.example.com/login
|
||||
# ... complete login flow ...
|
||||
|
||||
# All subsequent runs: already authenticated
|
||||
agent-browser --profile ~/.myapp-profile open https://app.example.com/dashboard
|
||||
```
|
||||
|
||||
Use different paths for different projects or test users:
|
||||
|
||||
```bash
|
||||
agent-browser --profile ~/.profiles/admin open https://app.example.com
|
||||
agent-browser --profile ~/.profiles/viewer open https://app.example.com
|
||||
```
|
||||
|
||||
Or set via environment variable:
|
||||
|
||||
```bash
|
||||
export AGENT_BROWSER_PROFILE=~/.myapp-profile
|
||||
agent-browser open https://app.example.com/dashboard
|
||||
```
|
||||
|
||||
## Session Persistence
|
||||
|
||||
Use `--session-name` to auto-save and restore cookies + localStorage by name, without managing files:
|
||||
|
||||
```bash
|
||||
# Auto-saves state on close, auto-restores on next launch
|
||||
agent-browser --session-name twitter open https://twitter.com
|
||||
# ... login flow ...
|
||||
agent-browser close # state saved to ~/.agent-browser/sessions/
|
||||
|
||||
# Next time: state is automatically restored
|
||||
agent-browser --session-name twitter open https://twitter.com
|
||||
```
|
||||
|
||||
Encrypt state at rest:
|
||||
|
||||
```bash
|
||||
export AGENT_BROWSER_ENCRYPTION_KEY=$(openssl rand -hex 32)
|
||||
agent-browser --session-name secure open https://app.example.com
|
||||
```
|
||||
|
||||
## Basic Login Flow
|
||||
|
||||
```bash
|
||||
# Navigate to login page
|
||||
agent-browser open https://app.example.com/login
|
||||
agent-browser wait --load networkidle
|
||||
|
||||
# Get form elements
|
||||
agent-browser snapshot -i
|
||||
# Output: @e1 [input type="email"], @e2 [input type="password"], @e3 [button] "Sign In"
|
||||
|
||||
# Fill credentials
|
||||
agent-browser fill @e1 "user@example.com"
|
||||
agent-browser fill @e2 "password123"
|
||||
|
||||
# Submit
|
||||
agent-browser click @e3
|
||||
agent-browser wait --load networkidle
|
||||
|
||||
# Verify login succeeded
|
||||
agent-browser get url # Should be dashboard, not login
|
||||
```
|
||||
|
||||
## Saving Authentication State
|
||||
|
||||
After logging in, save state for reuse:
|
||||
|
||||
```bash
|
||||
# Login first (see above)
|
||||
agent-browser open https://app.example.com/login
|
||||
agent-browser snapshot -i
|
||||
agent-browser fill @e1 "user@example.com"
|
||||
agent-browser fill @e2 "password123"
|
||||
agent-browser click @e3
|
||||
agent-browser wait --url "**/dashboard"
|
||||
|
||||
# Save authenticated state
|
||||
agent-browser state save ./auth-state.json
|
||||
```
|
||||
|
||||
## Restoring Authentication
|
||||
|
||||
Skip login by loading saved state:
|
||||
|
||||
```bash
|
||||
# Load saved auth state
|
||||
agent-browser state load ./auth-state.json
|
||||
|
||||
# Navigate directly to protected page
|
||||
agent-browser open https://app.example.com/dashboard
|
||||
|
||||
# Verify authenticated
|
||||
agent-browser snapshot -i
|
||||
```
|
||||
|
||||
## OAuth / SSO Flows
|
||||
|
||||
For OAuth redirects:
|
||||
|
||||
```bash
|
||||
# Start OAuth flow
|
||||
agent-browser open https://app.example.com/auth/google
|
||||
|
||||
# Handle redirects automatically
|
||||
agent-browser wait --url "**/accounts.google.com**"
|
||||
agent-browser snapshot -i
|
||||
|
||||
# Fill Google credentials
|
||||
agent-browser fill @e1 "user@gmail.com"
|
||||
agent-browser click @e2 # Next button
|
||||
agent-browser wait 2000
|
||||
agent-browser snapshot -i
|
||||
agent-browser fill @e3 "password"
|
||||
agent-browser click @e4 # Sign in
|
||||
|
||||
# Wait for redirect back
|
||||
agent-browser wait --url "**/app.example.com**"
|
||||
agent-browser state save ./oauth-state.json
|
||||
```
|
||||
|
||||
## Two-Factor Authentication
|
||||
|
||||
Handle 2FA with manual intervention:
|
||||
|
||||
```bash
|
||||
# Login with credentials
|
||||
agent-browser open https://app.example.com/login --headed # Show browser
|
||||
agent-browser snapshot -i
|
||||
agent-browser fill @e1 "user@example.com"
|
||||
agent-browser fill @e2 "password123"
|
||||
agent-browser click @e3
|
||||
|
||||
# Wait for user to complete 2FA manually
|
||||
echo "Complete 2FA in the browser window..."
|
||||
agent-browser wait --url "**/dashboard" --timeout 120000
|
||||
|
||||
# Save state after 2FA
|
||||
agent-browser state save ./2fa-state.json
|
||||
```
|
||||
|
||||
## HTTP Basic Auth
|
||||
|
||||
For sites using HTTP Basic Authentication:
|
||||
|
||||
```bash
|
||||
# Set credentials before navigation
|
||||
agent-browser set credentials username password
|
||||
|
||||
# Navigate to protected resource
|
||||
agent-browser open https://protected.example.com/api
|
||||
```
|
||||
|
||||
## Cookie-Based Auth
|
||||
|
||||
Manually set authentication cookies:
|
||||
|
||||
```bash
|
||||
# Set auth cookie
|
||||
agent-browser cookies set session_token "abc123xyz"
|
||||
|
||||
# Navigate to protected page
|
||||
agent-browser open https://app.example.com/dashboard
|
||||
```
|
||||
|
||||
## Token Refresh Handling
|
||||
|
||||
For sessions with expiring tokens:
|
||||
|
||||
```bash
|
||||
#!/bin/bash
|
||||
# Wrapper that handles token refresh
|
||||
|
||||
STATE_FILE="./auth-state.json"
|
||||
|
||||
# Try loading existing state
|
||||
if [[ -f "$STATE_FILE" ]]; then
|
||||
agent-browser state load "$STATE_FILE"
|
||||
agent-browser open https://app.example.com/dashboard
|
||||
|
||||
# Check if session is still valid
|
||||
URL=$(agent-browser get url)
|
||||
if [[ "$URL" == *"/login"* ]]; then
|
||||
echo "Session expired, re-authenticating..."
|
||||
# Perform fresh login
|
||||
agent-browser snapshot -i
|
||||
agent-browser fill @e1 "$USERNAME"
|
||||
agent-browser fill @e2 "$PASSWORD"
|
||||
agent-browser click @e3
|
||||
agent-browser wait --url "**/dashboard"
|
||||
agent-browser state save "$STATE_FILE"
|
||||
fi
|
||||
else
|
||||
# First-time login
|
||||
agent-browser open https://app.example.com/login
|
||||
# ... login flow ...
|
||||
fi
|
||||
```
|
||||
|
||||
## Security Best Practices
|
||||
|
||||
1. **Never commit state files** - They contain session tokens
|
||||
```bash
|
||||
echo "*.auth-state.json" >> .gitignore
|
||||
```
|
||||
|
||||
2. **Use environment variables for credentials**
|
||||
```bash
|
||||
agent-browser fill @e1 "$APP_USERNAME"
|
||||
agent-browser fill @e2 "$APP_PASSWORD"
|
||||
```
|
||||
|
||||
3. **Clean up after automation**
|
||||
```bash
|
||||
agent-browser cookies clear
|
||||
rm -f ./auth-state.json
|
||||
```
|
||||
|
||||
4. **Use short-lived sessions for CI/CD**
|
||||
```bash
|
||||
# Don't persist state in CI
|
||||
agent-browser open https://app.example.com/login
|
||||
# ... login and perform actions ...
|
||||
agent-browser close # Session ends, nothing persisted
|
||||
```
|
||||
@@ -0,0 +1,266 @@
|
||||
# Command Reference
|
||||
|
||||
Complete reference for all agent-browser commands. For quick start and common patterns, see SKILL.md.
|
||||
|
||||
## Navigation
|
||||
|
||||
```bash
|
||||
agent-browser open <url> # Navigate to URL (aliases: goto, navigate)
|
||||
# Supports: https://, http://, file://, about:, data://
|
||||
# Auto-prepends https:// if no protocol given
|
||||
agent-browser back # Go back
|
||||
agent-browser forward # Go forward
|
||||
agent-browser reload # Reload page
|
||||
agent-browser close # Close browser (aliases: quit, exit)
|
||||
agent-browser connect 9222 # Connect to browser via CDP port
|
||||
```
|
||||
|
||||
## Snapshot (page analysis)
|
||||
|
||||
```bash
|
||||
agent-browser snapshot # Full accessibility tree
|
||||
agent-browser snapshot -i # Interactive elements only (recommended)
|
||||
agent-browser snapshot -c # Compact output
|
||||
agent-browser snapshot -d 3 # Limit depth to 3
|
||||
agent-browser snapshot -s "#main" # Scope to CSS selector
|
||||
```
|
||||
|
||||
## Interactions (use @refs from snapshot)
|
||||
|
||||
```bash
|
||||
agent-browser click @e1 # Click
|
||||
agent-browser click @e1 --new-tab # Click and open in new tab
|
||||
agent-browser dblclick @e1 # Double-click
|
||||
agent-browser focus @e1 # Focus element
|
||||
agent-browser fill @e2 "text" # Clear and type
|
||||
agent-browser type @e2 "text" # Type without clearing
|
||||
agent-browser press Enter # Press key (alias: key)
|
||||
agent-browser press Control+a # Key combination
|
||||
agent-browser keydown Shift # Hold key down
|
||||
agent-browser keyup Shift # Release key
|
||||
agent-browser hover @e1 # Hover
|
||||
agent-browser check @e1 # Check checkbox
|
||||
agent-browser uncheck @e1 # Uncheck checkbox
|
||||
agent-browser select @e1 "value" # Select dropdown option
|
||||
agent-browser select @e1 "a" "b" # Select multiple options
|
||||
agent-browser scroll down 500 # Scroll page (default: down 300px)
|
||||
agent-browser scrollintoview @e1 # Scroll element into view (alias: scrollinto)
|
||||
agent-browser drag @e1 @e2 # Drag and drop
|
||||
agent-browser upload @e1 file.pdf # Upload files
|
||||
```
|
||||
|
||||
## Get Information
|
||||
|
||||
```bash
|
||||
agent-browser get text @e1 # Get element text
|
||||
agent-browser get html @e1 # Get innerHTML
|
||||
agent-browser get value @e1 # Get input value
|
||||
agent-browser get attr @e1 href # Get attribute
|
||||
agent-browser get title # Get page title
|
||||
agent-browser get url # Get current URL
|
||||
agent-browser get cdp-url # Get CDP WebSocket URL
|
||||
agent-browser get count ".item" # Count matching elements
|
||||
agent-browser get box @e1 # Get bounding box
|
||||
agent-browser get styles @e1 # Get computed styles (font, color, bg, etc.)
|
||||
```
|
||||
|
||||
## Check State
|
||||
|
||||
```bash
|
||||
agent-browser is visible @e1 # Check if visible
|
||||
agent-browser is enabled @e1 # Check if enabled
|
||||
agent-browser is checked @e1 # Check if checked
|
||||
```
|
||||
|
||||
## Screenshots and PDF
|
||||
|
||||
```bash
|
||||
agent-browser screenshot # Save to temporary directory
|
||||
agent-browser screenshot path.png # Save to specific path
|
||||
agent-browser screenshot --full # Full page
|
||||
agent-browser pdf output.pdf # Save as PDF
|
||||
```
|
||||
|
||||
## Video Recording
|
||||
|
||||
```bash
|
||||
agent-browser record start ./demo.webm # Start recording
|
||||
agent-browser click @e1 # Perform actions
|
||||
agent-browser record stop # Stop and save video
|
||||
agent-browser record restart ./take2.webm # Stop current + start new
|
||||
```
|
||||
|
||||
## Wait
|
||||
|
||||
```bash
|
||||
agent-browser wait @e1 # Wait for element
|
||||
agent-browser wait 2000 # Wait milliseconds
|
||||
agent-browser wait --text "Success" # Wait for text (or -t)
|
||||
agent-browser wait --url "**/dashboard" # Wait for URL pattern (or -u)
|
||||
agent-browser wait --load networkidle # Wait for network idle (or -l)
|
||||
agent-browser wait --fn "window.ready" # Wait for JS condition (or -f)
|
||||
```
|
||||
|
||||
## Mouse Control
|
||||
|
||||
```bash
|
||||
agent-browser mouse move 100 200 # Move mouse
|
||||
agent-browser mouse down left # Press button
|
||||
agent-browser mouse up left # Release button
|
||||
agent-browser mouse wheel 100 # Scroll wheel
|
||||
```
|
||||
|
||||
## Semantic Locators (alternative to refs)
|
||||
|
||||
```bash
|
||||
agent-browser find role button click --name "Submit"
|
||||
agent-browser find text "Sign In" click
|
||||
agent-browser find text "Sign In" click --exact # Exact match only
|
||||
agent-browser find label "Email" fill "user@test.com"
|
||||
agent-browser find placeholder "Search" type "query"
|
||||
agent-browser find alt "Logo" click
|
||||
agent-browser find title "Close" click
|
||||
agent-browser find testid "submit-btn" click
|
||||
agent-browser find first ".item" click
|
||||
agent-browser find last ".item" click
|
||||
agent-browser find nth 2 "a" hover
|
||||
```
|
||||
|
||||
## Browser Settings
|
||||
|
||||
```bash
|
||||
agent-browser set viewport 1920 1080 # Set viewport size
|
||||
agent-browser set viewport 1920 1080 2 # 2x retina (same CSS size, higher res screenshots)
|
||||
agent-browser set device "iPhone 14" # Emulate device
|
||||
agent-browser set geo 37.7749 -122.4194 # Set geolocation (alias: geolocation)
|
||||
agent-browser set offline on # Toggle offline mode
|
||||
agent-browser set headers '{"X-Key":"v"}' # Extra HTTP headers
|
||||
agent-browser set credentials user pass # HTTP basic auth (alias: auth)
|
||||
agent-browser set media dark # Emulate color scheme
|
||||
agent-browser set media light reduced-motion # Light mode + reduced motion
|
||||
```
|
||||
|
||||
## Cookies and Storage
|
||||
|
||||
```bash
|
||||
agent-browser cookies # Get all cookies
|
||||
agent-browser cookies set name value # Set cookie
|
||||
agent-browser cookies clear # Clear cookies
|
||||
agent-browser storage local # Get all localStorage
|
||||
agent-browser storage local key # Get specific key
|
||||
agent-browser storage local set k v # Set value
|
||||
agent-browser storage local clear # Clear all
|
||||
```
|
||||
|
||||
## Network
|
||||
|
||||
```bash
|
||||
agent-browser network route <url> # Intercept requests
|
||||
agent-browser network route <url> --abort # Block requests
|
||||
agent-browser network route <url> --body '{}' # Mock response
|
||||
agent-browser network unroute [url] # Remove routes
|
||||
agent-browser network requests # View tracked requests
|
||||
agent-browser network requests --filter api # Filter requests
|
||||
```
|
||||
|
||||
## Tabs and Windows
|
||||
|
||||
```bash
|
||||
agent-browser tab # List tabs
|
||||
agent-browser tab new [url] # New tab
|
||||
agent-browser tab 2 # Switch to tab by index
|
||||
agent-browser tab close # Close current tab
|
||||
agent-browser tab close 2 # Close tab by index
|
||||
agent-browser window new # New window
|
||||
```
|
||||
|
||||
## Frames
|
||||
|
||||
```bash
|
||||
agent-browser frame "#iframe" # Switch to iframe
|
||||
agent-browser frame main # Back to main frame
|
||||
```
|
||||
|
||||
## Dialogs
|
||||
|
||||
```bash
|
||||
agent-browser dialog accept [text] # Accept dialog
|
||||
agent-browser dialog dismiss # Dismiss dialog
|
||||
```
|
||||
|
||||
## JavaScript
|
||||
|
||||
```bash
|
||||
agent-browser eval "document.title" # Simple expressions only
|
||||
agent-browser eval -b "<base64>" # Any JavaScript (base64 encoded)
|
||||
agent-browser eval --stdin # Read script from stdin
|
||||
```
|
||||
|
||||
Use `-b`/`--base64` or `--stdin` for reliable execution. Shell escaping with nested quotes and special characters is error-prone.
|
||||
|
||||
```bash
|
||||
# Base64 encode your script, then:
|
||||
agent-browser eval -b "ZG9jdW1lbnQucXVlcnlTZWxlY3RvcignW3NyYyo9Il9uZXh0Il0nKQ=="
|
||||
|
||||
# Or use stdin with heredoc for multiline scripts:
|
||||
cat <<'EOF' | agent-browser eval --stdin
|
||||
const links = document.querySelectorAll('a');
|
||||
Array.from(links).map(a => a.href);
|
||||
EOF
|
||||
```
|
||||
|
||||
## State Management
|
||||
|
||||
```bash
|
||||
agent-browser state save auth.json # Save cookies, storage, auth state
|
||||
agent-browser state load auth.json # Restore saved state
|
||||
```
|
||||
|
||||
## Global Options
|
||||
|
||||
```bash
|
||||
agent-browser --session <name> ... # Isolated browser session
|
||||
agent-browser --json ... # JSON output for parsing
|
||||
agent-browser --headed ... # Show browser window (not headless)
|
||||
agent-browser --full ... # Full page screenshot (-f)
|
||||
agent-browser --cdp <port> ... # Connect via Chrome DevTools Protocol
|
||||
agent-browser -p <provider> ... # Cloud browser provider (--provider)
|
||||
agent-browser --proxy <url> ... # Use proxy server
|
||||
agent-browser --proxy-bypass <hosts> # Hosts to bypass proxy
|
||||
agent-browser --headers <json> ... # HTTP headers scoped to URL's origin
|
||||
agent-browser --executable-path <p> # Custom browser executable
|
||||
agent-browser --extension <path> ... # Load browser extension (repeatable)
|
||||
agent-browser --ignore-https-errors # Ignore SSL certificate errors
|
||||
agent-browser --help # Show help (-h)
|
||||
agent-browser --version # Show version (-V)
|
||||
agent-browser <command> --help # Show detailed help for a command
|
||||
```
|
||||
|
||||
## Debugging
|
||||
|
||||
```bash
|
||||
agent-browser --headed open example.com # Show browser window
|
||||
agent-browser --cdp 9222 snapshot # Connect via CDP port
|
||||
agent-browser connect 9222 # Alternative: connect command
|
||||
agent-browser console # View console messages
|
||||
agent-browser console --clear # Clear console
|
||||
agent-browser errors # View page errors
|
||||
agent-browser errors --clear # Clear errors
|
||||
agent-browser highlight @e1 # Highlight element
|
||||
agent-browser inspect # Open Chrome DevTools for this session
|
||||
agent-browser trace start # Start recording trace
|
||||
agent-browser trace stop trace.zip # Stop and save trace
|
||||
agent-browser profiler start # Start Chrome DevTools profiling
|
||||
agent-browser profiler stop trace.json # Stop and save profile
|
||||
```
|
||||
|
||||
## Environment Variables
|
||||
|
||||
```bash
|
||||
AGENT_BROWSER_SESSION="mysession" # Default session name
|
||||
AGENT_BROWSER_EXECUTABLE_PATH="/path/chrome" # Custom browser path
|
||||
AGENT_BROWSER_EXTENSIONS="/ext1,/ext2" # Comma-separated extension paths
|
||||
AGENT_BROWSER_PROVIDER="browserbase" # Cloud browser provider
|
||||
AGENT_BROWSER_STREAM_PORT="9223" # WebSocket streaming port
|
||||
AGENT_BROWSER_HOME="/path/to/agent-browser" # Custom install location
|
||||
```
|
||||
@@ -0,0 +1,120 @@
|
||||
# Profiling
|
||||
|
||||
Capture Chrome DevTools performance profiles during browser automation for performance analysis.
|
||||
|
||||
**Related**: [commands.md](commands.md) for full command reference, [SKILL.md](../SKILL.md) for quick start.
|
||||
|
||||
## Contents
|
||||
|
||||
- [Basic Profiling](#basic-profiling)
|
||||
- [Profiler Commands](#profiler-commands)
|
||||
- [Categories](#categories)
|
||||
- [Use Cases](#use-cases)
|
||||
- [Output Format](#output-format)
|
||||
- [Viewing Profiles](#viewing-profiles)
|
||||
- [Limitations](#limitations)
|
||||
|
||||
## Basic Profiling
|
||||
|
||||
```bash
|
||||
# Start profiling
|
||||
agent-browser profiler start
|
||||
|
||||
# Perform actions
|
||||
agent-browser navigate https://example.com
|
||||
agent-browser click "#button"
|
||||
agent-browser wait 1000
|
||||
|
||||
# Stop and save
|
||||
agent-browser profiler stop ./trace.json
|
||||
```
|
||||
|
||||
## Profiler Commands
|
||||
|
||||
```bash
|
||||
# Start profiling with default categories
|
||||
agent-browser profiler start
|
||||
|
||||
# Start with custom trace categories
|
||||
agent-browser profiler start --categories "devtools.timeline,v8.execute,blink.user_timing"
|
||||
|
||||
# Stop profiling and save to file
|
||||
agent-browser profiler stop ./trace.json
|
||||
```
|
||||
|
||||
## Categories
|
||||
|
||||
The `--categories` flag accepts a comma-separated list of Chrome trace categories. Default categories include:
|
||||
|
||||
- `devtools.timeline` -- standard DevTools performance traces
|
||||
- `v8.execute` -- time spent running JavaScript
|
||||
- `blink` -- renderer events
|
||||
- `blink.user_timing` -- `performance.mark()` / `performance.measure()` calls
|
||||
- `latencyInfo` -- input-to-latency tracking
|
||||
- `renderer.scheduler` -- task scheduling and execution
|
||||
- `toplevel` -- broad-spectrum basic events
|
||||
|
||||
Several `disabled-by-default-*` categories are also included for detailed timeline, call stack, and V8 CPU profiling data.
|
||||
|
||||
## Use Cases
|
||||
|
||||
### Diagnosing Slow Page Loads
|
||||
|
||||
```bash
|
||||
agent-browser profiler start
|
||||
agent-browser navigate https://app.example.com
|
||||
agent-browser wait --load networkidle
|
||||
agent-browser profiler stop ./page-load-profile.json
|
||||
```
|
||||
|
||||
### Profiling User Interactions
|
||||
|
||||
```bash
|
||||
agent-browser navigate https://app.example.com
|
||||
agent-browser profiler start
|
||||
agent-browser click "#submit"
|
||||
agent-browser wait 2000
|
||||
agent-browser profiler stop ./interaction-profile.json
|
||||
```
|
||||
|
||||
### CI Performance Regression Checks
|
||||
|
||||
```bash
|
||||
#!/bin/bash
|
||||
agent-browser profiler start
|
||||
agent-browser navigate https://app.example.com
|
||||
agent-browser wait --load networkidle
|
||||
agent-browser profiler stop "./profiles/build-${BUILD_ID}.json"
|
||||
```
|
||||
|
||||
## Output Format
|
||||
|
||||
The output is a JSON file in Chrome Trace Event format:
|
||||
|
||||
```json
|
||||
{
|
||||
"traceEvents": [
|
||||
{ "cat": "devtools.timeline", "name": "RunTask", "ph": "X", "ts": 12345, "dur": 100 },
|
||||
...
|
||||
],
|
||||
"metadata": {
|
||||
"clock-domain": "LINUX_CLOCK_MONOTONIC"
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
The `metadata.clock-domain` field is set based on the host platform (Linux or macOS). On Windows it is omitted.
|
||||
|
||||
## Viewing Profiles
|
||||
|
||||
Load the output JSON file in any of these tools:
|
||||
|
||||
- **Chrome DevTools**: Performance panel > Load profile (Ctrl+Shift+I > Performance)
|
||||
- **Perfetto UI**: https://ui.perfetto.dev/ -- drag and drop the JSON file
|
||||
- **Trace Viewer**: `chrome://tracing` in any Chromium browser
|
||||
|
||||
## Limitations
|
||||
|
||||
- Only works with Chromium-based browsers (Chrome, Edge). Not supported on Firefox or WebKit.
|
||||
- Trace data accumulates in memory while profiling is active (capped at 5 million events). Stop profiling promptly after the area of interest.
|
||||
- Data collection on stop has a 30-second timeout. If the browser is unresponsive, the stop command may fail.
|
||||
@@ -0,0 +1,194 @@
|
||||
# Proxy Support
|
||||
|
||||
Proxy configuration for geo-testing, rate limiting avoidance, and corporate environments.
|
||||
|
||||
**Related**: [commands.md](commands.md) for global options, [SKILL.md](../SKILL.md) for quick start.
|
||||
|
||||
## Contents
|
||||
|
||||
- [Basic Proxy Configuration](#basic-proxy-configuration)
|
||||
- [Authenticated Proxy](#authenticated-proxy)
|
||||
- [SOCKS Proxy](#socks-proxy)
|
||||
- [Proxy Bypass](#proxy-bypass)
|
||||
- [Common Use Cases](#common-use-cases)
|
||||
- [Verifying Proxy Connection](#verifying-proxy-connection)
|
||||
- [Troubleshooting](#troubleshooting)
|
||||
- [Best Practices](#best-practices)
|
||||
|
||||
## Basic Proxy Configuration
|
||||
|
||||
Use the `--proxy` flag or set proxy via environment variable:
|
||||
|
||||
```bash
|
||||
# Via CLI flag
|
||||
agent-browser --proxy "http://proxy.example.com:8080" open https://example.com
|
||||
|
||||
# Via environment variable
|
||||
export HTTP_PROXY="http://proxy.example.com:8080"
|
||||
agent-browser open https://example.com
|
||||
|
||||
# HTTPS proxy
|
||||
export HTTPS_PROXY="https://proxy.example.com:8080"
|
||||
agent-browser open https://example.com
|
||||
|
||||
# Both
|
||||
export HTTP_PROXY="http://proxy.example.com:8080"
|
||||
export HTTPS_PROXY="http://proxy.example.com:8080"
|
||||
agent-browser open https://example.com
|
||||
```
|
||||
|
||||
## Authenticated Proxy
|
||||
|
||||
For proxies requiring authentication:
|
||||
|
||||
```bash
|
||||
# Include credentials in URL
|
||||
export HTTP_PROXY="http://username:password@proxy.example.com:8080"
|
||||
agent-browser open https://example.com
|
||||
```
|
||||
|
||||
## SOCKS Proxy
|
||||
|
||||
```bash
|
||||
# SOCKS5 proxy
|
||||
export ALL_PROXY="socks5://proxy.example.com:1080"
|
||||
agent-browser open https://example.com
|
||||
|
||||
# SOCKS5 with auth
|
||||
export ALL_PROXY="socks5://user:pass@proxy.example.com:1080"
|
||||
agent-browser open https://example.com
|
||||
```
|
||||
|
||||
## Proxy Bypass
|
||||
|
||||
Skip proxy for specific domains using `--proxy-bypass` or `NO_PROXY`:
|
||||
|
||||
```bash
|
||||
# Via CLI flag
|
||||
agent-browser --proxy "http://proxy.example.com:8080" --proxy-bypass "localhost,*.internal.com" open https://example.com
|
||||
|
||||
# Via environment variable
|
||||
export NO_PROXY="localhost,127.0.0.1,.internal.company.com"
|
||||
agent-browser open https://internal.company.com # Direct connection
|
||||
agent-browser open https://external.com # Via proxy
|
||||
```
|
||||
|
||||
## Common Use Cases
|
||||
|
||||
### Geo-Location Testing
|
||||
|
||||
```bash
|
||||
#!/bin/bash
|
||||
# Test site from different regions using geo-located proxies
|
||||
|
||||
PROXIES=(
|
||||
"http://us-proxy.example.com:8080"
|
||||
"http://eu-proxy.example.com:8080"
|
||||
"http://asia-proxy.example.com:8080"
|
||||
)
|
||||
|
||||
for proxy in "${PROXIES[@]}"; do
|
||||
export HTTP_PROXY="$proxy"
|
||||
export HTTPS_PROXY="$proxy"
|
||||
|
||||
region=$(echo "$proxy" | grep -oP '^\w+-\w+')
|
||||
echo "Testing from: $region"
|
||||
|
||||
agent-browser --session "$region" open https://example.com
|
||||
agent-browser --session "$region" screenshot "./screenshots/$region.png"
|
||||
agent-browser --session "$region" close
|
||||
done
|
||||
```
|
||||
|
||||
### Rotating Proxies for Scraping
|
||||
|
||||
```bash
|
||||
#!/bin/bash
|
||||
# Rotate through proxy list to avoid rate limiting
|
||||
|
||||
PROXY_LIST=(
|
||||
"http://proxy1.example.com:8080"
|
||||
"http://proxy2.example.com:8080"
|
||||
"http://proxy3.example.com:8080"
|
||||
)
|
||||
|
||||
URLS=(
|
||||
"https://site.com/page1"
|
||||
"https://site.com/page2"
|
||||
"https://site.com/page3"
|
||||
)
|
||||
|
||||
for i in "${!URLS[@]}"; do
|
||||
proxy_index=$((i % ${#PROXY_LIST[@]}))
|
||||
export HTTP_PROXY="${PROXY_LIST[$proxy_index]}"
|
||||
export HTTPS_PROXY="${PROXY_LIST[$proxy_index]}"
|
||||
|
||||
agent-browser open "${URLS[$i]}"
|
||||
agent-browser get text body > "output-$i.txt"
|
||||
agent-browser close
|
||||
|
||||
sleep 1 # Polite delay
|
||||
done
|
||||
```
|
||||
|
||||
### Corporate Network Access
|
||||
|
||||
```bash
|
||||
#!/bin/bash
|
||||
# Access internal sites via corporate proxy
|
||||
|
||||
export HTTP_PROXY="http://corpproxy.company.com:8080"
|
||||
export HTTPS_PROXY="http://corpproxy.company.com:8080"
|
||||
export NO_PROXY="localhost,127.0.0.1,.company.com"
|
||||
|
||||
# External sites go through proxy
|
||||
agent-browser open https://external-vendor.com
|
||||
|
||||
# Internal sites bypass proxy
|
||||
agent-browser open https://intranet.company.com
|
||||
```
|
||||
|
||||
## Verifying Proxy Connection
|
||||
|
||||
```bash
|
||||
# Check your apparent IP
|
||||
agent-browser open https://httpbin.org/ip
|
||||
agent-browser get text body
|
||||
# Should show proxy's IP, not your real IP
|
||||
```
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
### Proxy Connection Failed
|
||||
|
||||
```bash
|
||||
# Test proxy connectivity first
|
||||
curl -x http://proxy.example.com:8080 https://httpbin.org/ip
|
||||
|
||||
# Check if proxy requires auth
|
||||
export HTTP_PROXY="http://user:pass@proxy.example.com:8080"
|
||||
```
|
||||
|
||||
### SSL/TLS Errors Through Proxy
|
||||
|
||||
Some proxies perform SSL inspection. If you encounter certificate errors:
|
||||
|
||||
```bash
|
||||
# For testing only - not recommended for production
|
||||
agent-browser open https://example.com --ignore-https-errors
|
||||
```
|
||||
|
||||
### Slow Performance
|
||||
|
||||
```bash
|
||||
# Use proxy only when necessary
|
||||
export NO_PROXY="*.cdn.com,*.static.com" # Direct CDN access
|
||||
```
|
||||
|
||||
## Best Practices
|
||||
|
||||
1. **Use environment variables** - Don't hardcode proxy credentials
|
||||
2. **Set NO_PROXY appropriately** - Avoid routing local traffic through proxy
|
||||
3. **Test proxy before automation** - Verify connectivity with simple requests
|
||||
4. **Handle proxy failures gracefully** - Implement retry logic for unstable proxies
|
||||
5. **Rotate proxies for large scraping jobs** - Distribute load and avoid bans
|
||||
@@ -0,0 +1,193 @@
|
||||
# Session Management
|
||||
|
||||
Multiple isolated browser sessions with state persistence and concurrent browsing.
|
||||
|
||||
**Related**: [authentication.md](authentication.md) for login patterns, [SKILL.md](../SKILL.md) for quick start.
|
||||
|
||||
## Contents
|
||||
|
||||
- [Named Sessions](#named-sessions)
|
||||
- [Session Isolation Properties](#session-isolation-properties)
|
||||
- [Session State Persistence](#session-state-persistence)
|
||||
- [Common Patterns](#common-patterns)
|
||||
- [Default Session](#default-session)
|
||||
- [Session Cleanup](#session-cleanup)
|
||||
- [Best Practices](#best-practices)
|
||||
|
||||
## Named Sessions
|
||||
|
||||
Use `--session` flag to isolate browser contexts:
|
||||
|
||||
```bash
|
||||
# Session 1: Authentication flow
|
||||
agent-browser --session auth open https://app.example.com/login
|
||||
|
||||
# Session 2: Public browsing (separate cookies, storage)
|
||||
agent-browser --session public open https://example.com
|
||||
|
||||
# Commands are isolated by session
|
||||
agent-browser --session auth fill @e1 "user@example.com"
|
||||
agent-browser --session public get text body
|
||||
```
|
||||
|
||||
## Session Isolation Properties
|
||||
|
||||
Each session has independent:
|
||||
- Cookies
|
||||
- LocalStorage / SessionStorage
|
||||
- IndexedDB
|
||||
- Cache
|
||||
- Browsing history
|
||||
- Open tabs
|
||||
|
||||
## Session State Persistence
|
||||
|
||||
### Save Session State
|
||||
|
||||
```bash
|
||||
# Save cookies, storage, and auth state
|
||||
agent-browser state save /path/to/auth-state.json
|
||||
```
|
||||
|
||||
### Load Session State
|
||||
|
||||
```bash
|
||||
# Restore saved state
|
||||
agent-browser state load /path/to/auth-state.json
|
||||
|
||||
# Continue with authenticated session
|
||||
agent-browser open https://app.example.com/dashboard
|
||||
```
|
||||
|
||||
### State File Contents
|
||||
|
||||
```json
|
||||
{
|
||||
"cookies": [...],
|
||||
"localStorage": {...},
|
||||
"sessionStorage": {...},
|
||||
"origins": [...]
|
||||
}
|
||||
```
|
||||
|
||||
## Common Patterns
|
||||
|
||||
### Authenticated Session Reuse
|
||||
|
||||
```bash
|
||||
#!/bin/bash
|
||||
# Save login state once, reuse many times
|
||||
|
||||
STATE_FILE="/tmp/auth-state.json"
|
||||
|
||||
# Check if we have saved state
|
||||
if [[ -f "$STATE_FILE" ]]; then
|
||||
agent-browser state load "$STATE_FILE"
|
||||
agent-browser open https://app.example.com/dashboard
|
||||
else
|
||||
# Perform login
|
||||
agent-browser open https://app.example.com/login
|
||||
agent-browser snapshot -i
|
||||
agent-browser fill @e1 "$USERNAME"
|
||||
agent-browser fill @e2 "$PASSWORD"
|
||||
agent-browser click @e3
|
||||
agent-browser wait --load networkidle
|
||||
|
||||
# Save for future use
|
||||
agent-browser state save "$STATE_FILE"
|
||||
fi
|
||||
```
|
||||
|
||||
### Concurrent Scraping
|
||||
|
||||
```bash
|
||||
#!/bin/bash
|
||||
# Scrape multiple sites concurrently
|
||||
|
||||
# Start all sessions
|
||||
agent-browser --session site1 open https://site1.com &
|
||||
agent-browser --session site2 open https://site2.com &
|
||||
agent-browser --session site3 open https://site3.com &
|
||||
wait
|
||||
|
||||
# Extract from each
|
||||
agent-browser --session site1 get text body > site1.txt
|
||||
agent-browser --session site2 get text body > site2.txt
|
||||
agent-browser --session site3 get text body > site3.txt
|
||||
|
||||
# Cleanup
|
||||
agent-browser --session site1 close
|
||||
agent-browser --session site2 close
|
||||
agent-browser --session site3 close
|
||||
```
|
||||
|
||||
### A/B Testing Sessions
|
||||
|
||||
```bash
|
||||
# Test different user experiences
|
||||
agent-browser --session variant-a open "https://app.com?variant=a"
|
||||
agent-browser --session variant-b open "https://app.com?variant=b"
|
||||
|
||||
# Compare
|
||||
agent-browser --session variant-a screenshot /tmp/variant-a.png
|
||||
agent-browser --session variant-b screenshot /tmp/variant-b.png
|
||||
```
|
||||
|
||||
## Default Session
|
||||
|
||||
When `--session` is omitted, commands use the default session:
|
||||
|
||||
```bash
|
||||
# These use the same default session
|
||||
agent-browser open https://example.com
|
||||
agent-browser snapshot -i
|
||||
agent-browser close # Closes default session
|
||||
```
|
||||
|
||||
## Session Cleanup
|
||||
|
||||
```bash
|
||||
# Close specific session
|
||||
agent-browser --session auth close
|
||||
|
||||
# List active sessions
|
||||
agent-browser session list
|
||||
```
|
||||
|
||||
## Best Practices
|
||||
|
||||
### 1. Name Sessions Semantically
|
||||
|
||||
```bash
|
||||
# GOOD: Clear purpose
|
||||
agent-browser --session github-auth open https://github.com
|
||||
agent-browser --session docs-scrape open https://docs.example.com
|
||||
|
||||
# AVOID: Generic names
|
||||
agent-browser --session s1 open https://github.com
|
||||
```
|
||||
|
||||
### 2. Always Clean Up
|
||||
|
||||
```bash
|
||||
# Close sessions when done
|
||||
agent-browser --session auth close
|
||||
agent-browser --session scrape close
|
||||
```
|
||||
|
||||
### 3. Handle State Files Securely
|
||||
|
||||
```bash
|
||||
# Don't commit state files (contain auth tokens!)
|
||||
echo "*.auth-state.json" >> .gitignore
|
||||
|
||||
# Delete after use
|
||||
rm /tmp/auth-state.json
|
||||
```
|
||||
|
||||
### 4. Timeout Long Sessions
|
||||
|
||||
```bash
|
||||
# Set timeout for automated scripts
|
||||
timeout 60 agent-browser --session long-task get text body
|
||||
```
|
||||
@@ -0,0 +1,194 @@
|
||||
# Snapshot and Refs
|
||||
|
||||
Compact element references that reduce context usage dramatically for AI agents.
|
||||
|
||||
**Related**: [commands.md](commands.md) for full command reference, [SKILL.md](../SKILL.md) for quick start.
|
||||
|
||||
## Contents
|
||||
|
||||
- [How Refs Work](#how-refs-work)
|
||||
- [Snapshot Command](#the-snapshot-command)
|
||||
- [Using Refs](#using-refs)
|
||||
- [Ref Lifecycle](#ref-lifecycle)
|
||||
- [Best Practices](#best-practices)
|
||||
- [Ref Notation Details](#ref-notation-details)
|
||||
- [Troubleshooting](#troubleshooting)
|
||||
|
||||
## How Refs Work
|
||||
|
||||
Traditional approach:
|
||||
```
|
||||
Full DOM/HTML -> AI parses -> CSS selector -> Action (~3000-5000 tokens)
|
||||
```
|
||||
|
||||
agent-browser approach:
|
||||
```
|
||||
Compact snapshot -> @refs assigned -> Direct interaction (~200-400 tokens)
|
||||
```
|
||||
|
||||
## The Snapshot Command
|
||||
|
||||
```bash
|
||||
# Basic snapshot (shows page structure)
|
||||
agent-browser snapshot
|
||||
|
||||
# Interactive snapshot (-i flag) - RECOMMENDED
|
||||
agent-browser snapshot -i
|
||||
```
|
||||
|
||||
### Snapshot Output Format
|
||||
|
||||
```
|
||||
Page: Example Site - Home
|
||||
URL: https://example.com
|
||||
|
||||
@e1 [header]
|
||||
@e2 [nav]
|
||||
@e3 [a] "Home"
|
||||
@e4 [a] "Products"
|
||||
@e5 [a] "About"
|
||||
@e6 [button] "Sign In"
|
||||
|
||||
@e7 [main]
|
||||
@e8 [h1] "Welcome"
|
||||
@e9 [form]
|
||||
@e10 [input type="email"] placeholder="Email"
|
||||
@e11 [input type="password"] placeholder="Password"
|
||||
@e12 [button type="submit"] "Log In"
|
||||
|
||||
@e13 [footer]
|
||||
@e14 [a] "Privacy Policy"
|
||||
```
|
||||
|
||||
## Using Refs
|
||||
|
||||
Once you have refs, interact directly:
|
||||
|
||||
```bash
|
||||
# Click the "Sign In" button
|
||||
agent-browser click @e6
|
||||
|
||||
# Fill email input
|
||||
agent-browser fill @e10 "user@example.com"
|
||||
|
||||
# Fill password
|
||||
agent-browser fill @e11 "password123"
|
||||
|
||||
# Submit the form
|
||||
agent-browser click @e12
|
||||
```
|
||||
|
||||
## Ref Lifecycle
|
||||
|
||||
**IMPORTANT**: Refs are invalidated when the page changes!
|
||||
|
||||
```bash
|
||||
# Get initial snapshot
|
||||
agent-browser snapshot -i
|
||||
# @e1 [button] "Next"
|
||||
|
||||
# Click triggers page change
|
||||
agent-browser click @e1
|
||||
|
||||
# MUST re-snapshot to get new refs!
|
||||
agent-browser snapshot -i
|
||||
# @e1 [h1] "Page 2" <- Different element now!
|
||||
```
|
||||
|
||||
## Best Practices
|
||||
|
||||
### 1. Always Snapshot Before Interacting
|
||||
|
||||
```bash
|
||||
# CORRECT
|
||||
agent-browser open https://example.com
|
||||
agent-browser snapshot -i # Get refs first
|
||||
agent-browser click @e1 # Use ref
|
||||
|
||||
# WRONG
|
||||
agent-browser open https://example.com
|
||||
agent-browser click @e1 # Ref doesn't exist yet!
|
||||
```
|
||||
|
||||
### 2. Re-Snapshot After Navigation
|
||||
|
||||
```bash
|
||||
agent-browser click @e5 # Navigates to new page
|
||||
agent-browser snapshot -i # Get new refs
|
||||
agent-browser click @e1 # Use new refs
|
||||
```
|
||||
|
||||
### 3. Re-Snapshot After Dynamic Changes
|
||||
|
||||
```bash
|
||||
agent-browser click @e1 # Opens dropdown
|
||||
agent-browser snapshot -i # See dropdown items
|
||||
agent-browser click @e7 # Select item
|
||||
```
|
||||
|
||||
### 4. Snapshot Specific Regions
|
||||
|
||||
For complex pages, snapshot specific areas:
|
||||
|
||||
```bash
|
||||
# Snapshot just the form
|
||||
agent-browser snapshot @e9
|
||||
```
|
||||
|
||||
## Ref Notation Details
|
||||
|
||||
```
|
||||
@e1 [tag type="value"] "text content" placeholder="hint"
|
||||
| | | | |
|
||||
| | | | +- Additional attributes
|
||||
| | | +- Visible text
|
||||
| | +- Key attributes shown
|
||||
| +- HTML tag name
|
||||
+- Unique ref ID
|
||||
```
|
||||
|
||||
### Common Patterns
|
||||
|
||||
```
|
||||
@e1 [button] "Submit" # Button with text
|
||||
@e2 [input type="email"] # Email input
|
||||
@e3 [input type="password"] # Password input
|
||||
@e4 [a href="/page"] "Link Text" # Anchor link
|
||||
@e5 [select] # Dropdown
|
||||
@e6 [textarea] placeholder="Message" # Text area
|
||||
@e7 [div class="modal"] # Container (when relevant)
|
||||
@e8 [img alt="Logo"] # Image
|
||||
@e9 [checkbox] checked # Checked checkbox
|
||||
@e10 [radio] selected # Selected radio
|
||||
```
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
### "Ref not found" Error
|
||||
|
||||
```bash
|
||||
# Ref may have changed - re-snapshot
|
||||
agent-browser snapshot -i
|
||||
```
|
||||
|
||||
### Element Not Visible in Snapshot
|
||||
|
||||
```bash
|
||||
# Scroll down to reveal element
|
||||
agent-browser scroll down 1000
|
||||
agent-browser snapshot -i
|
||||
|
||||
# Or wait for dynamic content
|
||||
agent-browser wait 1000
|
||||
agent-browser snapshot -i
|
||||
```
|
||||
|
||||
### Too Many Elements
|
||||
|
||||
```bash
|
||||
# Snapshot specific container
|
||||
agent-browser snapshot @e5
|
||||
|
||||
# Or use get text for content-only extraction
|
||||
agent-browser get text @e5
|
||||
```
|
||||
@@ -0,0 +1,173 @@
|
||||
# Video Recording
|
||||
|
||||
Capture browser automation as video for debugging, documentation, or verification.
|
||||
|
||||
**Related**: [commands.md](commands.md) for full command reference, [SKILL.md](../SKILL.md) for quick start.
|
||||
|
||||
## Contents
|
||||
|
||||
- [Basic Recording](#basic-recording)
|
||||
- [Recording Commands](#recording-commands)
|
||||
- [Use Cases](#use-cases)
|
||||
- [Best Practices](#best-practices)
|
||||
- [Output Format](#output-format)
|
||||
- [Limitations](#limitations)
|
||||
|
||||
## Basic Recording
|
||||
|
||||
```bash
|
||||
# Start recording
|
||||
agent-browser record start ./demo.webm
|
||||
|
||||
# Perform actions
|
||||
agent-browser open https://example.com
|
||||
agent-browser snapshot -i
|
||||
agent-browser click @e1
|
||||
agent-browser fill @e2 "test input"
|
||||
|
||||
# Stop and save
|
||||
agent-browser record stop
|
||||
```
|
||||
|
||||
## Recording Commands
|
||||
|
||||
```bash
|
||||
# Start recording to file
|
||||
agent-browser record start ./output.webm
|
||||
|
||||
# Stop current recording
|
||||
agent-browser record stop
|
||||
|
||||
# Restart with new file (stops current + starts new)
|
||||
agent-browser record restart ./take2.webm
|
||||
```
|
||||
|
||||
## Use Cases
|
||||
|
||||
### Debugging Failed Automation
|
||||
|
||||
```bash
|
||||
#!/bin/bash
|
||||
# Record automation for debugging
|
||||
|
||||
agent-browser record start ./debug-$(date +%Y%m%d-%H%M%S).webm
|
||||
|
||||
# Run your automation
|
||||
agent-browser open https://app.example.com
|
||||
agent-browser snapshot -i
|
||||
agent-browser click @e1 || {
|
||||
echo "Click failed - check recording"
|
||||
agent-browser record stop
|
||||
exit 1
|
||||
}
|
||||
|
||||
agent-browser record stop
|
||||
```
|
||||
|
||||
### Documentation Generation
|
||||
|
||||
```bash
|
||||
#!/bin/bash
|
||||
# Record workflow for documentation
|
||||
|
||||
agent-browser record start ./docs/how-to-login.webm
|
||||
|
||||
agent-browser open https://app.example.com/login
|
||||
agent-browser wait 1000 # Pause for visibility
|
||||
|
||||
agent-browser snapshot -i
|
||||
agent-browser fill @e1 "demo@example.com"
|
||||
agent-browser wait 500
|
||||
|
||||
agent-browser fill @e2 "password"
|
||||
agent-browser wait 500
|
||||
|
||||
agent-browser click @e3
|
||||
agent-browser wait --load networkidle
|
||||
agent-browser wait 1000 # Show result
|
||||
|
||||
agent-browser record stop
|
||||
```
|
||||
|
||||
### CI/CD Test Evidence
|
||||
|
||||
```bash
|
||||
#!/bin/bash
|
||||
# Record E2E test runs for CI artifacts
|
||||
|
||||
TEST_NAME="${1:-e2e-test}"
|
||||
RECORDING_DIR="./test-recordings"
|
||||
mkdir -p "$RECORDING_DIR"
|
||||
|
||||
agent-browser record start "$RECORDING_DIR/$TEST_NAME-$(date +%s).webm"
|
||||
|
||||
# Run test
|
||||
if run_e2e_test; then
|
||||
echo "Test passed"
|
||||
else
|
||||
echo "Test failed - recording saved"
|
||||
fi
|
||||
|
||||
agent-browser record stop
|
||||
```
|
||||
|
||||
## Best Practices
|
||||
|
||||
### 1. Add Pauses for Clarity
|
||||
|
||||
```bash
|
||||
# Slow down for human viewing
|
||||
agent-browser click @e1
|
||||
agent-browser wait 500 # Let viewer see result
|
||||
```
|
||||
|
||||
### 2. Use Descriptive Filenames
|
||||
|
||||
```bash
|
||||
# Include context in filename
|
||||
agent-browser record start ./recordings/login-flow-2024-01-15.webm
|
||||
agent-browser record start ./recordings/checkout-test-run-42.webm
|
||||
```
|
||||
|
||||
### 3. Handle Recording in Error Cases
|
||||
|
||||
```bash
|
||||
#!/bin/bash
|
||||
set -e
|
||||
|
||||
cleanup() {
|
||||
agent-browser record stop 2>/dev/null || true
|
||||
agent-browser close 2>/dev/null || true
|
||||
}
|
||||
trap cleanup EXIT
|
||||
|
||||
agent-browser record start ./automation.webm
|
||||
# ... automation steps ...
|
||||
```
|
||||
|
||||
### 4. Combine with Screenshots
|
||||
|
||||
```bash
|
||||
# Record video AND capture key frames
|
||||
agent-browser record start ./flow.webm
|
||||
|
||||
agent-browser open https://example.com
|
||||
agent-browser screenshot ./screenshots/step1-homepage.png
|
||||
|
||||
agent-browser click @e1
|
||||
agent-browser screenshot ./screenshots/step2-after-click.png
|
||||
|
||||
agent-browser record stop
|
||||
```
|
||||
|
||||
## Output Format
|
||||
|
||||
- Default format: WebM (VP8/VP9 codec)
|
||||
- Compatible with all modern browsers and video players
|
||||
- Compressed but high quality
|
||||
|
||||
## Limitations
|
||||
|
||||
- Recording adds slight overhead to automation
|
||||
- Large recordings can consume significant disk space
|
||||
- Some headless environments may have codec limitations
|
||||
@@ -0,0 +1,105 @@
|
||||
#!/bin/bash
|
||||
# Template: Authenticated Session Workflow
|
||||
# Purpose: Login once, save state, reuse for subsequent runs
|
||||
# Usage: ./authenticated-session.sh <login-url> [state-file]
|
||||
#
|
||||
# RECOMMENDED: Use the auth vault instead of this template:
|
||||
# echo "<pass>" | agent-browser auth save myapp --url <login-url> --username <user> --password-stdin
|
||||
# agent-browser auth login myapp
|
||||
# The auth vault stores credentials securely and the LLM never sees passwords.
|
||||
#
|
||||
# Environment variables:
|
||||
# APP_USERNAME - Login username/email
|
||||
# APP_PASSWORD - Login password
|
||||
#
|
||||
# Two modes:
|
||||
# 1. Discovery mode (default): Shows form structure so you can identify refs
|
||||
# 2. Login mode: Performs actual login after you update the refs
|
||||
#
|
||||
# Setup steps:
|
||||
# 1. Run once to see form structure (discovery mode)
|
||||
# 2. Update refs in LOGIN FLOW section below
|
||||
# 3. Set APP_USERNAME and APP_PASSWORD
|
||||
# 4. Delete the DISCOVERY section
|
||||
|
||||
set -euo pipefail
|
||||
|
||||
LOGIN_URL="${1:?Usage: $0 <login-url> [state-file]}"
|
||||
STATE_FILE="${2:-./auth-state.json}"
|
||||
|
||||
echo "Authentication workflow: $LOGIN_URL"
|
||||
|
||||
# ================================================================
|
||||
# SAVED STATE: Skip login if valid saved state exists
|
||||
# ================================================================
|
||||
if [[ -f "$STATE_FILE" ]]; then
|
||||
echo "Loading saved state from $STATE_FILE..."
|
||||
if agent-browser --state "$STATE_FILE" open "$LOGIN_URL" 2>/dev/null; then
|
||||
agent-browser wait --load networkidle
|
||||
|
||||
CURRENT_URL=$(agent-browser get url)
|
||||
if [[ "$CURRENT_URL" != *"login"* ]] && [[ "$CURRENT_URL" != *"signin"* ]]; then
|
||||
echo "Session restored successfully"
|
||||
agent-browser snapshot -i
|
||||
exit 0
|
||||
fi
|
||||
echo "Session expired, performing fresh login..."
|
||||
agent-browser close 2>/dev/null || true
|
||||
else
|
||||
echo "Failed to load state, re-authenticating..."
|
||||
fi
|
||||
rm -f "$STATE_FILE"
|
||||
fi
|
||||
|
||||
# ================================================================
|
||||
# DISCOVERY MODE: Shows form structure (delete after setup)
|
||||
# ================================================================
|
||||
echo "Opening login page..."
|
||||
agent-browser open "$LOGIN_URL"
|
||||
agent-browser wait --load networkidle
|
||||
|
||||
echo ""
|
||||
echo "Login form structure:"
|
||||
echo "---"
|
||||
agent-browser snapshot -i
|
||||
echo "---"
|
||||
echo ""
|
||||
echo "Next steps:"
|
||||
echo " 1. Note the refs: username=@e?, password=@e?, submit=@e?"
|
||||
echo " 2. Update the LOGIN FLOW section below with your refs"
|
||||
echo " 3. Set: export APP_USERNAME='...' APP_PASSWORD='...'"
|
||||
echo " 4. Delete this DISCOVERY MODE section"
|
||||
echo ""
|
||||
agent-browser close
|
||||
exit 0
|
||||
|
||||
# ================================================================
|
||||
# LOGIN FLOW: Uncomment and customize after discovery
|
||||
# ================================================================
|
||||
# : "${APP_USERNAME:?Set APP_USERNAME environment variable}"
|
||||
# : "${APP_PASSWORD:?Set APP_PASSWORD environment variable}"
|
||||
#
|
||||
# agent-browser open "$LOGIN_URL"
|
||||
# agent-browser wait --load networkidle
|
||||
# agent-browser snapshot -i
|
||||
#
|
||||
# # Fill credentials (update refs to match your form)
|
||||
# agent-browser fill @e1 "$APP_USERNAME"
|
||||
# agent-browser fill @e2 "$APP_PASSWORD"
|
||||
# agent-browser click @e3
|
||||
# agent-browser wait --load networkidle
|
||||
#
|
||||
# # Verify login succeeded
|
||||
# FINAL_URL=$(agent-browser get url)
|
||||
# if [[ "$FINAL_URL" == *"login"* ]] || [[ "$FINAL_URL" == *"signin"* ]]; then
|
||||
# echo "Login failed - still on login page"
|
||||
# agent-browser screenshot /tmp/login-failed.png
|
||||
# agent-browser close
|
||||
# exit 1
|
||||
# fi
|
||||
#
|
||||
# # Save state for future runs
|
||||
# echo "Saving state to $STATE_FILE"
|
||||
# agent-browser state save "$STATE_FILE"
|
||||
# echo "Login successful"
|
||||
# agent-browser snapshot -i
|
||||
@@ -0,0 +1,69 @@
|
||||
#!/bin/bash
|
||||
# Template: Content Capture Workflow
|
||||
# Purpose: Extract content from web pages (text, screenshots, PDF)
|
||||
# Usage: ./capture-workflow.sh <url> [output-dir]
|
||||
#
|
||||
# Outputs:
|
||||
# - page-full.png: Full page screenshot
|
||||
# - page-structure.txt: Page element structure with refs
|
||||
# - page-text.txt: All text content
|
||||
# - page.pdf: PDF version
|
||||
#
|
||||
# Optional: Load auth state for protected pages
|
||||
|
||||
set -euo pipefail
|
||||
|
||||
TARGET_URL="${1:?Usage: $0 <url> [output-dir]}"
|
||||
OUTPUT_DIR="${2:-.}"
|
||||
|
||||
echo "Capturing: $TARGET_URL"
|
||||
mkdir -p "$OUTPUT_DIR"
|
||||
|
||||
# Optional: Load authentication state
|
||||
# if [[ -f "./auth-state.json" ]]; then
|
||||
# echo "Loading authentication state..."
|
||||
# agent-browser state load "./auth-state.json"
|
||||
# fi
|
||||
|
||||
# Navigate to target
|
||||
agent-browser open "$TARGET_URL"
|
||||
agent-browser wait --load networkidle
|
||||
|
||||
# Get metadata
|
||||
TITLE=$(agent-browser get title)
|
||||
URL=$(agent-browser get url)
|
||||
echo "Title: $TITLE"
|
||||
echo "URL: $URL"
|
||||
|
||||
# Capture full page screenshot
|
||||
agent-browser screenshot --full "$OUTPUT_DIR/page-full.png"
|
||||
echo "Saved: $OUTPUT_DIR/page-full.png"
|
||||
|
||||
# Get page structure with refs
|
||||
agent-browser snapshot -i > "$OUTPUT_DIR/page-structure.txt"
|
||||
echo "Saved: $OUTPUT_DIR/page-structure.txt"
|
||||
|
||||
# Extract all text content
|
||||
agent-browser get text body > "$OUTPUT_DIR/page-text.txt"
|
||||
echo "Saved: $OUTPUT_DIR/page-text.txt"
|
||||
|
||||
# Save as PDF
|
||||
agent-browser pdf "$OUTPUT_DIR/page.pdf"
|
||||
echo "Saved: $OUTPUT_DIR/page.pdf"
|
||||
|
||||
# Optional: Extract specific elements using refs from structure
|
||||
# agent-browser get text @e5 > "$OUTPUT_DIR/main-content.txt"
|
||||
|
||||
# Optional: Handle infinite scroll pages
|
||||
# for i in {1..5}; do
|
||||
# agent-browser scroll down 1000
|
||||
# agent-browser wait 1000
|
||||
# done
|
||||
# agent-browser screenshot --full "$OUTPUT_DIR/page-scrolled.png"
|
||||
|
||||
# Cleanup
|
||||
agent-browser close
|
||||
|
||||
echo ""
|
||||
echo "Capture complete:"
|
||||
ls -la "$OUTPUT_DIR"
|
||||
@@ -0,0 +1,62 @@
|
||||
#!/bin/bash
|
||||
# Template: Form Automation Workflow
|
||||
# Purpose: Fill and submit web forms with validation
|
||||
# Usage: ./form-automation.sh <form-url>
|
||||
#
|
||||
# This template demonstrates the snapshot-interact-verify pattern:
|
||||
# 1. Navigate to form
|
||||
# 2. Snapshot to get element refs
|
||||
# 3. Fill fields using refs
|
||||
# 4. Submit and verify result
|
||||
#
|
||||
# Customize: Update the refs (@e1, @e2, etc.) based on your form's snapshot output
|
||||
|
||||
set -euo pipefail
|
||||
|
||||
FORM_URL="${1:?Usage: $0 <form-url>}"
|
||||
|
||||
echo "Form automation: $FORM_URL"
|
||||
|
||||
# Step 1: Navigate to form
|
||||
agent-browser open "$FORM_URL"
|
||||
agent-browser wait --load networkidle
|
||||
|
||||
# Step 2: Snapshot to discover form elements
|
||||
echo ""
|
||||
echo "Form structure:"
|
||||
agent-browser snapshot -i
|
||||
|
||||
# Step 3: Fill form fields (customize these refs based on snapshot output)
|
||||
#
|
||||
# Common field types:
|
||||
# agent-browser fill @e1 "John Doe" # Text input
|
||||
# agent-browser fill @e2 "user@example.com" # Email input
|
||||
# agent-browser fill @e3 "SecureP@ss123" # Password input
|
||||
# agent-browser select @e4 "Option Value" # Dropdown
|
||||
# agent-browser check @e5 # Checkbox
|
||||
# agent-browser click @e6 # Radio button
|
||||
# agent-browser fill @e7 "Multi-line text" # Textarea
|
||||
# agent-browser upload @e8 /path/to/file.pdf # File upload
|
||||
#
|
||||
# Uncomment and modify:
|
||||
# agent-browser fill @e1 "Test User"
|
||||
# agent-browser fill @e2 "test@example.com"
|
||||
# agent-browser click @e3 # Submit button
|
||||
|
||||
# Step 4: Wait for submission
|
||||
# agent-browser wait --load networkidle
|
||||
# agent-browser wait --url "**/success" # Or wait for redirect
|
||||
|
||||
# Step 5: Verify result
|
||||
echo ""
|
||||
echo "Result:"
|
||||
agent-browser get url
|
||||
agent-browser snapshot -i
|
||||
|
||||
# Optional: Capture evidence
|
||||
agent-browser screenshot /tmp/form-result.png
|
||||
echo "Screenshot saved: /tmp/form-result.png"
|
||||
|
||||
# Cleanup
|
||||
agent-browser close
|
||||
echo "Done"
|
||||
@@ -1,190 +0,0 @@
|
||||
---
|
||||
name: brainstorming
|
||||
description: This skill should be used before implementing features, building components, or making changes. It guides exploring user intent, approaches, and design decisions before planning. Triggers on "let's brainstorm", "help me think through", "what should we build", "explore approaches", ambiguous feature requests, or when the user's request has multiple valid interpretations that need clarification.
|
||||
---
|
||||
|
||||
# Brainstorming
|
||||
|
||||
This skill provides detailed process knowledge for effective brainstorming sessions that clarify **WHAT** to build before diving into **HOW** to build it.
|
||||
|
||||
## When to Use This Skill
|
||||
|
||||
Brainstorming is valuable when:
|
||||
- Requirements are unclear or ambiguous
|
||||
- Multiple approaches could solve the problem
|
||||
- Trade-offs need to be explored with the user
|
||||
- The user hasn't fully articulated what they want
|
||||
- The feature scope needs refinement
|
||||
|
||||
Brainstorming can be skipped when:
|
||||
- Requirements are explicit and detailed
|
||||
- The user knows exactly what they want
|
||||
- The task is a straightforward bug fix or well-defined change
|
||||
|
||||
## Core Process
|
||||
|
||||
### Phase 0: Assess Requirement Clarity
|
||||
|
||||
Before diving into questions, assess whether brainstorming is needed.
|
||||
|
||||
**Signals that requirements are clear:**
|
||||
- User provided specific acceptance criteria
|
||||
- User referenced existing patterns to follow
|
||||
- User described exact behavior expected
|
||||
- Scope is constrained and well-defined
|
||||
|
||||
**Signals that brainstorming is needed:**
|
||||
- User used vague terms ("make it better", "add something like")
|
||||
- Multiple reasonable interpretations exist
|
||||
- Trade-offs haven't been discussed
|
||||
- User seems unsure about the approach
|
||||
|
||||
If requirements are clear, suggest: "Your requirements seem clear. Consider proceeding directly to planning or implementation."
|
||||
|
||||
### Phase 1: Understand the Idea
|
||||
|
||||
Ask questions **one at a time** to understand the user's intent. Avoid overwhelming with multiple questions.
|
||||
|
||||
**Question Techniques:**
|
||||
|
||||
1. **Prefer multiple choice when natural options exist**
|
||||
- Good: "Should the notification be: (a) email only, (b) in-app only, or (c) both?"
|
||||
- Avoid: "How should users be notified?"
|
||||
|
||||
2. **Start broad, then narrow**
|
||||
- First: What is the core purpose?
|
||||
- Then: Who are the users?
|
||||
- Finally: What constraints exist?
|
||||
|
||||
3. **Validate assumptions explicitly**
|
||||
- "I'm assuming users will be logged in. Is that correct?"
|
||||
|
||||
4. **Ask about success criteria early**
|
||||
- "How will you know this feature is working well?"
|
||||
|
||||
**Key Topics to Explore:**
|
||||
|
||||
| Topic | Example Questions |
|
||||
|-------|-------------------|
|
||||
| Purpose | What problem does this solve? What's the motivation? |
|
||||
| Users | Who uses this? What's their context? |
|
||||
| Constraints | Any technical limitations? Timeline? Dependencies? |
|
||||
| Success | How will you measure success? What's the happy path? |
|
||||
| Edge Cases | What shouldn't happen? Any error states to consider? |
|
||||
| Existing Patterns | Are there similar features in the codebase to follow? |
|
||||
|
||||
**Exit Condition:** Continue until the idea is clear OR user says "proceed" or "let's move on"
|
||||
|
||||
### Phase 2: Explore Approaches
|
||||
|
||||
After understanding the idea, propose 2-3 concrete approaches.
|
||||
|
||||
**Structure for Each Approach:**
|
||||
|
||||
```markdown
|
||||
### Approach A: [Name]
|
||||
|
||||
[2-3 sentence description]
|
||||
|
||||
**Pros:**
|
||||
- [Benefit 1]
|
||||
- [Benefit 2]
|
||||
|
||||
**Cons:**
|
||||
- [Drawback 1]
|
||||
- [Drawback 2]
|
||||
|
||||
**Best when:** [Circumstances where this approach shines]
|
||||
```
|
||||
|
||||
**Guidelines:**
|
||||
- Lead with a recommendation and explain why
|
||||
- Be honest about trade-offs
|
||||
- Consider YAGNI—simpler is usually better
|
||||
- Reference codebase patterns when relevant
|
||||
|
||||
### Phase 3: Capture the Design
|
||||
|
||||
Summarize key decisions in a structured format.
|
||||
|
||||
**Design Doc Structure:**
|
||||
|
||||
```markdown
|
||||
---
|
||||
date: YYYY-MM-DD
|
||||
topic: <kebab-case-topic>
|
||||
---
|
||||
|
||||
# <Topic Title>
|
||||
|
||||
## What We're Building
|
||||
[Concise description—1-2 paragraphs max]
|
||||
|
||||
## Why This Approach
|
||||
[Brief explanation of approaches considered and why this one was chosen]
|
||||
|
||||
## Key Decisions
|
||||
- [Decision 1]: [Rationale]
|
||||
- [Decision 2]: [Rationale]
|
||||
|
||||
## Open Questions
|
||||
- [Any unresolved questions for the planning phase]
|
||||
|
||||
## Next Steps
|
||||
→ `/workflows:plan` for implementation details
|
||||
```
|
||||
|
||||
**Output Location:** `docs/brainstorms/YYYY-MM-DD-<topic>-brainstorm.md`
|
||||
|
||||
### Phase 4: Handoff
|
||||
|
||||
Present clear options for what to do next:
|
||||
|
||||
1. **Proceed to planning** → Run `/workflows:plan`
|
||||
2. **Refine further** → Continue exploring the design
|
||||
3. **Done for now** → User will return later
|
||||
|
||||
## YAGNI Principles
|
||||
|
||||
During brainstorming, actively resist complexity:
|
||||
|
||||
- **Don't design for hypothetical future requirements**
|
||||
- **Choose the simplest approach that solves the stated problem**
|
||||
- **Prefer boring, proven patterns over clever solutions**
|
||||
- **Ask "Do we really need this?" when complexity emerges**
|
||||
- **Defer decisions that don't need to be made now**
|
||||
|
||||
## Incremental Validation
|
||||
|
||||
Keep sections short—200-300 words maximum. After each section of output, pause to validate understanding:
|
||||
|
||||
- "Does this match what you had in mind?"
|
||||
- "Any adjustments before we continue?"
|
||||
- "Is this the direction you want to go?"
|
||||
|
||||
This prevents wasted effort on misaligned designs.
|
||||
|
||||
## Anti-Patterns to Avoid
|
||||
|
||||
| Anti-Pattern | Better Approach |
|
||||
|--------------|-----------------|
|
||||
| Asking 5 questions at once | Ask one at a time |
|
||||
| Jumping to implementation details | Stay focused on WHAT, not HOW |
|
||||
| Proposing overly complex solutions | Start simple, add complexity only if needed |
|
||||
| Ignoring existing codebase patterns | Research what exists first |
|
||||
| Making assumptions without validating | State assumptions explicitly and confirm |
|
||||
| Creating lengthy design documents | Keep it concise—details go in the plan |
|
||||
|
||||
## Integration with Planning
|
||||
|
||||
Brainstorming answers **WHAT** to build:
|
||||
- Requirements and acceptance criteria
|
||||
- Chosen approach and rationale
|
||||
- Key decisions and trade-offs
|
||||
|
||||
Planning answers **HOW** to build it:
|
||||
- Implementation steps and file changes
|
||||
- Technical details and code patterns
|
||||
- Testing strategy and verification
|
||||
|
||||
When brainstorm output exists, `/workflows:plan` should detect it and use it as input, skipping its own idea refinement phase.
|
||||
335
plugins/compound-engineering/skills/ce-brainstorm/SKILL.md
Normal file
335
plugins/compound-engineering/skills/ce-brainstorm/SKILL.md
Normal file
@@ -0,0 +1,335 @@
|
||||
---
|
||||
name: ce:brainstorm
|
||||
description: 'Explore requirements and approaches through collaborative dialogue before writing a right-sized requirements document and planning implementation. Use for feature ideas, problem framing, when the user says ''let''s brainstorm'', or when they want to think through options before deciding what to build. Also use when a user describes a vague or ambitious feature request, asks ''what should we build'', ''help me think through X'', presents a problem with multiple valid solutions, or seems unsure about scope or direction — even if they don''t explicitly ask to brainstorm.'
|
||||
argument-hint: "[feature idea or problem to explore]"
|
||||
---
|
||||
|
||||
# Brainstorm a Feature or Improvement
|
||||
|
||||
**Note: The current year is 2026.** Use this when dating requirements documents.
|
||||
|
||||
Brainstorming helps answer **WHAT** to build through collaborative dialogue. It precedes `/ce:plan`, which answers **HOW** to build it.
|
||||
|
||||
The durable output of this workflow is a **requirements document**. In other workflows this might be called a lightweight PRD or feature brief. In compound engineering, keep the workflow name `brainstorm`, but make the written artifact strong enough that planning does not need to invent product behavior, scope boundaries, or success criteria.
|
||||
|
||||
This skill does not implement code. It explores, clarifies, and documents decisions for later planning or execution.
|
||||
|
||||
## Core Principles
|
||||
|
||||
1. **Assess scope first** - Match the amount of ceremony to the size and ambiguity of the work.
|
||||
2. **Be a thinking partner** - Suggest alternatives, challenge assumptions, and explore what-ifs instead of only extracting requirements.
|
||||
3. **Resolve product decisions here** - User-facing behavior, scope boundaries, and success criteria belong in this workflow. Detailed implementation belongs in planning.
|
||||
4. **Keep implementation out of the requirements doc by default** - Do not include libraries, schemas, endpoints, file layouts, or code-level design unless the brainstorm itself is inherently about a technical or architectural change.
|
||||
5. **Right-size the artifact** - Simple work gets a compact requirements document or brief alignment. Larger work gets a fuller document. Do not add ceremony that does not help planning.
|
||||
6. **Apply YAGNI to carrying cost, not coding effort** - Prefer the simplest approach that delivers meaningful value. Avoid speculative complexity and hypothetical future-proofing, but low-cost polish or delight is worth including when its ongoing cost is small and easy to maintain.
|
||||
|
||||
## Interaction Rules
|
||||
|
||||
1. **Ask one question at a time** - Do not batch several unrelated questions into one message.
|
||||
2. **Prefer single-select multiple choice** - Use single-select when choosing one direction, one priority, or one next step.
|
||||
3. **Use multi-select rarely and intentionally** - Use it only for compatible sets such as goals, constraints, non-goals, or success criteria that can all coexist. If prioritization matters, follow up by asking which selected item is primary.
|
||||
4. **Use the platform's question tool when available** - When asking the user a question, prefer the platform's blocking question tool if one exists (`AskUserQuestion` in Claude Code, `request_user_input` in Codex, `ask_user` in Gemini). Otherwise, present numbered options in chat and wait for the user's reply before proceeding.
|
||||
|
||||
## Output Guidance
|
||||
|
||||
- **Keep outputs concise** - Prefer short sections, brief bullets, and only enough detail to support the next decision.
|
||||
|
||||
## Feature Description
|
||||
|
||||
<feature_description> #$ARGUMENTS </feature_description>
|
||||
|
||||
**If the feature description above is empty, ask the user:** "What would you like to explore? Please describe the feature, problem, or improvement you're thinking about."
|
||||
|
||||
Do not proceed until you have a feature description from the user.
|
||||
|
||||
## Execution Flow
|
||||
|
||||
### Phase 0: Resume, Assess, and Route
|
||||
|
||||
#### 0.1 Resume Existing Work When Appropriate
|
||||
|
||||
If the user references an existing brainstorm topic or document, or there is an obvious recent matching `*-requirements.md` file in `docs/brainstorms/`:
|
||||
- Read the document
|
||||
- Confirm with the user before resuming: "Found an existing requirements doc for [topic]. Should I continue from this, or start fresh?"
|
||||
- If resuming, summarize the current state briefly, continue from its existing decisions and outstanding questions, and update the existing document instead of creating a duplicate
|
||||
|
||||
#### 0.2 Assess Whether Brainstorming Is Needed
|
||||
|
||||
**Clear requirements indicators:**
|
||||
- Specific acceptance criteria provided
|
||||
- Referenced existing patterns to follow
|
||||
- Described exact expected behavior
|
||||
- Constrained, well-defined scope
|
||||
|
||||
**If requirements are already clear:**
|
||||
Keep the interaction brief. Confirm understanding and present concise next-step options rather than forcing a long brainstorm. Only write a short requirements document when a durable handoff to planning or later review would be valuable. Skip Phase 1.1 and 1.2 entirely — go straight to Phase 1.3 or Phase 3.
|
||||
|
||||
#### 0.3 Assess Scope
|
||||
|
||||
Use the feature description plus a light repo scan to classify the work:
|
||||
- **Lightweight** - small, well-bounded, low ambiguity
|
||||
- **Standard** - normal feature or bounded refactor with some decisions to make
|
||||
- **Deep** - cross-cutting, strategic, or highly ambiguous
|
||||
|
||||
If the scope is unclear, ask one targeted question to disambiguate and then proceed.
|
||||
|
||||
### Phase 1: Understand the Idea
|
||||
|
||||
#### 1.1 Existing Context Scan
|
||||
|
||||
Scan the repo before substantive brainstorming. Match depth to scope:
|
||||
|
||||
**Lightweight** — Search for the topic, check if something similar already exists, and move on.
|
||||
|
||||
**Standard and Deep** — Two passes:
|
||||
|
||||
*Constraint Check* — Check project instruction files (`AGENTS.md`, and `CLAUDE.md` only if retained as compatibility context) for workflow, product, or scope constraints that affect the brainstorm. If these add nothing, move on.
|
||||
|
||||
*Topic Scan* — Search for relevant terms. Read the most relevant existing artifact if one exists (brainstorm, plan, spec, skill, feature doc). Skim adjacent examples covering similar behavior.
|
||||
|
||||
If nothing obvious appears after a short scan, say so and continue. Do not drift into technical planning — avoid inspecting tests, migrations, deployment, or low-level architecture unless the brainstorm is itself about a technical decision.
|
||||
|
||||
#### 1.2 Product Pressure Test
|
||||
|
||||
Before generating approaches, challenge the request to catch misframing. Match depth to scope:
|
||||
|
||||
**Lightweight:**
|
||||
- Is this solving the real user problem?
|
||||
- Are we duplicating something that already covers this?
|
||||
- Is there a clearly better framing with near-zero extra cost?
|
||||
|
||||
**Standard:**
|
||||
- Is this the right problem, or a proxy for a more important one?
|
||||
- What user or business outcome actually matters here?
|
||||
- What happens if we do nothing?
|
||||
- Is there a nearby framing that creates more user value without more carrying cost? If so, what complexity does it add?
|
||||
- Given the current project state, user goal, and constraints, what is the single highest-leverage move right now: the request as framed, a reframing, one adjacent addition, a simplification, or doing nothing?
|
||||
- Favor moves that compound value, reduce future carrying cost, or make the product meaningfully more useful or compelling
|
||||
- Use the result to sharpen the conversation, not to bulldoze the user's intent
|
||||
|
||||
**Deep** — Standard questions plus:
|
||||
- What durable capability should this create in 6-12 months?
|
||||
- Does this move the product toward that, or is it only a local patch?
|
||||
|
||||
#### 1.3 Collaborative Dialogue
|
||||
|
||||
Use the platform's blocking question tool when available (see Interaction Rules). Otherwise, present numbered options in chat and wait for the user's reply before proceeding.
|
||||
|
||||
**Guidelines:**
|
||||
- Ask questions **one at a time**
|
||||
- Prefer multiple choice when natural options exist
|
||||
- Prefer **single-select** when choosing one direction, one priority, or one next step
|
||||
- Use **multi-select** only for compatible sets that can all coexist; if prioritization matters, ask which selected item is primary
|
||||
- Start broad (problem, users, value) then narrow (constraints, exclusions, edge cases)
|
||||
- Clarify the problem frame, validate assumptions, and ask about success criteria
|
||||
- Make requirements concrete enough that planning will not need to invent behavior
|
||||
- Surface dependencies or prerequisites only when they materially affect scope
|
||||
- Resolve product decisions here; leave technical implementation choices for planning
|
||||
- Bring ideas, alternatives, and challenges instead of only interviewing
|
||||
|
||||
**Exit condition:** Continue until the idea is clear OR the user explicitly wants to proceed.
|
||||
|
||||
### Phase 2: Explore Approaches
|
||||
|
||||
If multiple plausible directions remain, propose **2-3 concrete approaches** based on research and conversation. Otherwise state the recommended direction directly.
|
||||
|
||||
When useful, include one deliberately higher-upside alternative:
|
||||
- Identify what adjacent addition or reframing would most increase usefulness, compounding value, or durability without disproportionate carrying cost. Present it as a challenger option alongside the baseline, not as the default. Omit it when the work is already obviously over-scoped or the baseline request is clearly the right move.
|
||||
|
||||
For each approach, provide:
|
||||
- Brief description (2-3 sentences)
|
||||
- Pros and cons
|
||||
- Key risks or unknowns
|
||||
- When it's best suited
|
||||
|
||||
Lead with your recommendation and explain why. Prefer simpler solutions when added complexity creates real carrying cost, but do not reject low-cost, high-value polish just because it is not strictly necessary.
|
||||
|
||||
If one approach is clearly best and alternatives are not meaningful, skip the menu and state the recommendation directly.
|
||||
|
||||
If relevant, call out whether the choice is:
|
||||
- Reuse an existing pattern
|
||||
- Extend an existing capability
|
||||
- Build something net new
|
||||
|
||||
### Phase 3: Capture the Requirements
|
||||
|
||||
Write or update a requirements document only when the conversation produced durable decisions worth preserving.
|
||||
|
||||
This document should behave like a lightweight PRD without PRD ceremony. Include what planning needs to execute well, and skip sections that add no value for the scope.
|
||||
|
||||
The requirements document is for product definition and scope control. Do **not** include implementation details such as libraries, schemas, endpoints, file layouts, or code structure unless the brainstorm is inherently technical and those details are themselves the subject of the decision.
|
||||
|
||||
**Required content for non-trivial work:**
|
||||
- Problem frame
|
||||
- Concrete requirements or intended behavior with stable IDs
|
||||
- Scope boundaries
|
||||
- Success criteria
|
||||
|
||||
**Include when materially useful:**
|
||||
- Key decisions and rationale
|
||||
- Dependencies or assumptions
|
||||
- Outstanding questions
|
||||
- Alternatives considered
|
||||
- High-level technical direction only when the work is inherently technical and the direction is part of the product/architecture decision
|
||||
|
||||
**Document structure:** Use this template and omit clearly inapplicable optional sections:
|
||||
|
||||
```markdown
|
||||
---
|
||||
date: YYYY-MM-DD
|
||||
topic: <kebab-case-topic>
|
||||
---
|
||||
|
||||
# <Topic Title>
|
||||
|
||||
## Problem Frame
|
||||
[Who is affected, what is changing, and why it matters]
|
||||
|
||||
## Requirements
|
||||
- R1. [Concrete user-facing behavior or requirement]
|
||||
- R2. [Concrete user-facing behavior or requirement]
|
||||
|
||||
## Success Criteria
|
||||
- [How we will know this solved the right problem]
|
||||
|
||||
## Scope Boundaries
|
||||
- [Deliberate non-goal or exclusion]
|
||||
|
||||
## Key Decisions
|
||||
- [Decision]: [Rationale]
|
||||
|
||||
## Dependencies / Assumptions
|
||||
- [Only include if material]
|
||||
|
||||
## Outstanding Questions
|
||||
|
||||
### Resolve Before Planning
|
||||
- [Affects R1][User decision] [Question that must be answered before planning can proceed]
|
||||
|
||||
### Deferred to Planning
|
||||
- [Affects R2][Technical] [Question that should be answered during planning or codebase exploration]
|
||||
- [Affects R2][Needs research] [Question that likely requires research during planning]
|
||||
|
||||
## Next Steps
|
||||
[If `Resolve Before Planning` is empty: `→ /ce:plan` for structured implementation planning]
|
||||
[If `Resolve Before Planning` is not empty: `→ Resume /ce:brainstorm` to resolve blocking questions before planning]
|
||||
```
|
||||
|
||||
For **Standard** and **Deep** brainstorms, a requirements document is usually warranted.
|
||||
|
||||
For **Lightweight** brainstorms, keep the document compact. Skip document creation when the user only needs brief alignment and no durable decisions need to be preserved.
|
||||
|
||||
For very small requirements docs with only 1-3 simple requirements, plain bullet requirements are acceptable. For **Standard** and **Deep** requirements docs, use stable IDs like `R1`, `R2`, `R3` so planning and later review can refer to them unambiguously.
|
||||
|
||||
When the work is simple, combine sections rather than padding them. A short requirements document is better than a bloated one.
|
||||
|
||||
Before finalizing, check:
|
||||
- What would `ce:plan` still have to invent if this brainstorm ended now?
|
||||
- Do any requirements depend on something claimed to be out of scope?
|
||||
- Are any unresolved items actually product decisions rather than planning questions?
|
||||
- Did implementation details leak in when they shouldn't have?
|
||||
- Is there a low-cost change that would make this materially more useful?
|
||||
|
||||
If planning would need to invent product behavior, scope boundaries, or success criteria, the brainstorm is not complete yet.
|
||||
|
||||
Ensure `docs/brainstorms/` directory exists before writing.
|
||||
|
||||
If a document contains outstanding questions:
|
||||
- Use `Resolve Before Planning` only for questions that truly block planning
|
||||
- If `Resolve Before Planning` is non-empty, keep working those questions during the brainstorm by default
|
||||
- If the user explicitly wants to proceed anyway, convert each remaining item into an explicit decision, assumption, or `Deferred to Planning` question before proceeding
|
||||
- Do not force resolution of technical questions during brainstorming just to remove uncertainty
|
||||
- Put technical questions, or questions that require validation or research, under `Deferred to Planning` when they are better answered there
|
||||
- Use tags like `[Needs research]` when the planner should likely investigate the question rather than answer it from repo context alone
|
||||
- Carry deferred questions forward explicitly rather than treating them as a failure to finish the requirements doc
|
||||
|
||||
### Phase 4: Handoff
|
||||
|
||||
#### 4.1 Present Next-Step Options
|
||||
|
||||
Present next steps using the platform's blocking question tool when available (see Interaction Rules). Otherwise present numbered options in chat and end the turn.
|
||||
|
||||
If `Resolve Before Planning` contains any items:
|
||||
- Ask the blocking questions now, one at a time, by default
|
||||
- If the user explicitly wants to proceed anyway, first convert each remaining item into an explicit decision, assumption, or `Deferred to Planning` question
|
||||
- If the user chooses to pause instead, present the handoff as paused or blocked rather than complete
|
||||
- Do not offer `Proceed to planning` or `Proceed directly to work` while `Resolve Before Planning` remains non-empty
|
||||
|
||||
**Question when no blocking questions remain:** "Brainstorm complete. What would you like to do next?"
|
||||
|
||||
**Question when blocking questions remain and user wants to pause:** "Brainstorm paused. Planning is blocked until the remaining questions are resolved. What would you like to do next?"
|
||||
|
||||
Present only the options that apply:
|
||||
- **Proceed to planning (Recommended)** - Run `/ce:plan` for structured implementation planning
|
||||
- **Proceed directly to work** - Only offer this when scope is lightweight, success criteria are clear, scope boundaries are clear, and no meaningful technical or research questions remain
|
||||
- **Review and refine** - Offer this only when a requirements document exists and can be improved through structured review
|
||||
- **Ask more questions** - Continue clarifying scope, preferences, or edge cases
|
||||
- **Share to Proof** - Offer this only when a requirements document exists
|
||||
- **Done for now** - Return later
|
||||
|
||||
If the direct-to-work gate is not satisfied, omit that option entirely.
|
||||
|
||||
#### 4.2 Handle the Selected Option
|
||||
|
||||
**If user selects "Proceed to planning (Recommended)":**
|
||||
|
||||
Immediately run `/ce:plan` in the current session. Pass the requirements document path when one exists; otherwise pass a concise summary of the finalized brainstorm decisions. Do not print the closing summary first.
|
||||
|
||||
**If user selects "Proceed directly to work":**
|
||||
|
||||
Immediately run `/ce:work` in the current session using the finalized brainstorm output as context. If a compact requirements document exists, pass its path. Do not print the closing summary first.
|
||||
|
||||
**If user selects "Share to Proof":**
|
||||
|
||||
```bash
|
||||
CONTENT=$(cat docs/brainstorms/YYYY-MM-DD-<topic>-requirements.md)
|
||||
TITLE="Requirements: <topic title>"
|
||||
RESPONSE=$(curl -s -X POST https://www.proofeditor.ai/share/markdown \
|
||||
-H "Content-Type: application/json" \
|
||||
-d "$(jq -n --arg title "$TITLE" --arg markdown "$CONTENT" --arg by "ai:compound" '{title: $title, markdown: $markdown, by: $by}')")
|
||||
PROOF_URL=$(echo "$RESPONSE" | jq -r '.tokenUrl')
|
||||
```
|
||||
|
||||
Display the URL prominently: `View & collaborate in Proof: <PROOF_URL>`
|
||||
|
||||
If the curl fails, skip silently. Then return to the Phase 4 options.
|
||||
|
||||
**If user selects "Ask more questions":** Return to Phase 1.3 (Collaborative Dialogue) and continue asking the user questions one at a time to further refine the design. Probe deeper into edge cases, constraints, preferences, or areas not yet explored. Continue until the user is satisfied, then return to Phase 4. Do not show the closing summary yet.
|
||||
|
||||
**If user selects "Review and refine":**
|
||||
|
||||
Load the `document-review` skill and apply it to the requirements document.
|
||||
|
||||
When document-review returns "Review complete", return to the normal Phase 4 options and present only the options that still apply. Do not show the closing summary yet.
|
||||
|
||||
#### 4.3 Closing Summary
|
||||
|
||||
Use the closing summary only when this run of the workflow is ending or handing off, not when returning to the Phase 4 options.
|
||||
|
||||
When complete and ready for planning, display:
|
||||
|
||||
```text
|
||||
Brainstorm complete!
|
||||
|
||||
Requirements doc: docs/brainstorms/YYYY-MM-DD-<topic>-requirements.md # if one was created
|
||||
|
||||
Key decisions:
|
||||
- [Decision 1]
|
||||
- [Decision 2]
|
||||
|
||||
Recommended next step: `/ce:plan`
|
||||
```
|
||||
|
||||
If the user pauses with `Resolve Before Planning` still populated, display:
|
||||
|
||||
```text
|
||||
Brainstorm paused.
|
||||
|
||||
Requirements doc: docs/brainstorms/YYYY-MM-DD-<topic>-requirements.md # if one was created
|
||||
|
||||
Planning is blocked by:
|
||||
- [Blocking question 1]
|
||||
- [Blocking question 2]
|
||||
|
||||
Resume with `/ce:brainstorm` when ready to resolve these before planning.
|
||||
```
|
||||
527
plugins/compound-engineering/skills/ce-compound-refresh/SKILL.md
Normal file
527
plugins/compound-engineering/skills/ce-compound-refresh/SKILL.md
Normal file
@@ -0,0 +1,527 @@
|
||||
---
|
||||
name: ce:compound-refresh
|
||||
description: Refresh stale or drifting learnings and pattern docs in docs/solutions/ by reviewing, updating, replacing, or archiving them against the current codebase. Use after refactors, migrations, dependency upgrades, or when a retrieved learning feels outdated or wrong. Also use when reviewing docs/solutions/ for accuracy, when a recently solved problem contradicts an existing learning, or when pattern docs no longer reflect current code.
|
||||
argument-hint: "[mode:autonomous] [optional: scope hint]"
|
||||
disable-model-invocation: true
|
||||
---
|
||||
|
||||
# Compound Refresh
|
||||
|
||||
Maintain the quality of `docs/solutions/` over time. This workflow reviews existing learnings against the current codebase, then refreshes any derived pattern docs that depend on them.
|
||||
|
||||
## Mode Detection
|
||||
|
||||
Check if `$ARGUMENTS` contains `mode:autonomous`. If present, strip it from arguments (use the remainder as a scope hint) and run in **autonomous mode**.
|
||||
|
||||
| Mode | When | Behavior |
|
||||
|------|------|----------|
|
||||
| **Interactive** (default) | User is present and can answer questions | Ask for decisions on ambiguous cases, confirm actions |
|
||||
| **Autonomous** | `mode:autonomous` in arguments | No user interaction. Apply all unambiguous actions (Keep, Update, auto-Archive, Replace with sufficient evidence). Mark ambiguous cases as stale. Generate a summary report at the end. |
|
||||
|
||||
### Autonomous mode rules
|
||||
|
||||
- **Skip all user questions.** Never pause for input.
|
||||
- **Process all docs in scope.** No scope narrowing questions — if no scope hint was provided, process everything.
|
||||
- **Attempt all safe actions:** Keep (no-op), Update (fix references), auto-Archive (unambiguous criteria met), Replace (when evidence is sufficient). If a write succeeds, record it as **applied**. If a write fails (e.g., permission denied), record the action as **recommended** in the report and continue — do not stop or ask for permissions.
|
||||
- **Mark as stale when uncertain.** If classification is genuinely ambiguous (Update vs Replace vs Archive) or Replace evidence is insufficient, mark as stale with `status: stale`, `stale_reason`, and `stale_date` in the frontmatter. If even the stale-marking write fails, include it as a recommendation.
|
||||
- **Use conservative confidence.** In interactive mode, borderline cases get a user question. In autonomous mode, borderline cases get marked stale. Err toward stale-marking over incorrect action.
|
||||
- **Always generate a report.** The report is the primary deliverable. It has two sections: **Applied** (actions that were successfully written) and **Recommended** (actions that could not be written, with full rationale so a human can apply them or run the skill interactively). The report structure is the same regardless of what permissions were granted — the only difference is which section each action lands in.
|
||||
|
||||
## Interaction Principles
|
||||
|
||||
**These principles apply to interactive mode only. In autonomous mode, skip all user questions and apply the autonomous mode rules above.**
|
||||
|
||||
Follow the same interaction style as `ce:brainstorm`:
|
||||
|
||||
- Ask questions **one at a time** — use the platform's blocking question tool when available (`AskUserQuestion` in Claude Code, `request_user_input` in Codex, `ask_user` in Gemini). Otherwise, present numbered options in plain text and wait for the user's reply before continuing
|
||||
- Prefer **multiple choice** when natural options exist
|
||||
- Start with **scope and intent**, then narrow only when needed
|
||||
- Do **not** ask the user to make decisions before you have evidence
|
||||
- Lead with a recommendation and explain it briefly
|
||||
|
||||
The goal is not to force the user through a checklist. The goal is to help them make a good maintenance decision with the smallest amount of friction.
|
||||
|
||||
## Refresh Order
|
||||
|
||||
Refresh in this order:
|
||||
|
||||
1. Review the relevant individual learning docs first
|
||||
2. Note which learnings stayed valid, were updated, were replaced, or were archived
|
||||
3. Then review any pattern docs that depend on those learnings
|
||||
|
||||
Why this order:
|
||||
|
||||
- learning docs are the primary evidence
|
||||
- pattern docs are derived from one or more learnings
|
||||
- stale learnings can make a pattern look more valid than it really is
|
||||
|
||||
If the user starts by naming a pattern doc, you may begin there to understand the concern, but inspect the supporting learning docs before changing the pattern.
|
||||
|
||||
## Maintenance Model
|
||||
|
||||
For each candidate artifact, classify it into one of four outcomes:
|
||||
|
||||
| Outcome | Meaning | Default action |
|
||||
|---------|---------|----------------|
|
||||
| **Keep** | Still accurate and still useful | No file edit by default; report that it was reviewed and remains trustworthy |
|
||||
| **Update** | Core solution is still correct, but references drifted | Apply evidence-backed in-place edits |
|
||||
| **Replace** | The old artifact is now misleading, but there is a known better replacement | Create a trustworthy successor or revised pattern, then mark/archive the old artifact as needed |
|
||||
| **Archive** | No longer useful or applicable | Move the obsolete artifact to `docs/solutions/_archived/` with archive metadata when appropriate |
|
||||
|
||||
## Core Rules
|
||||
|
||||
1. **Evidence informs judgment.** The signals below are inputs, not a mechanical scorecard. Use engineering judgment to decide whether the artifact is still trustworthy.
|
||||
2. **Prefer no-write Keep.** Do not update a doc just to leave a review breadcrumb.
|
||||
3. **Match docs to reality, not the reverse.** When current code differs from a learning, update the learning to reflect the current code. The skill's job is doc accuracy, not code review — do not ask the user whether code changes were "intentional" or "a regression." If the code changed, the doc should match. If the user thinks the code is wrong, that is a separate concern outside this workflow.
|
||||
4. **Be decisive, minimize questions.** When evidence is clear (file renamed, class moved, reference broken), apply the update. In interactive mode, only ask the user when the right action is genuinely ambiguous. In autonomous mode, mark ambiguous cases as stale instead of asking. The goal is automated maintenance with human oversight on judgment calls, not a question for every finding.
|
||||
5. **Avoid low-value churn.** Do not edit a doc just to fix a typo, polish wording, or make cosmetic changes that do not materially improve accuracy or usability.
|
||||
6. **Use Update only for meaningful, evidence-backed drift.** Paths, module names, related links, category metadata, code snippets, and clearly stale wording are fair game when fixing them materially improves accuracy.
|
||||
7. **Use Replace only when there is a real replacement.** That means either:
|
||||
- the current conversation contains a recently solved, verified replacement fix, or
|
||||
- the user has provided enough concrete replacement context to document the successor honestly, or
|
||||
- the codebase investigation found the current approach and can document it as the successor, or
|
||||
- newer docs, pattern docs, PRs, or issues provide strong successor evidence.
|
||||
8. **Archive when the code is gone.** If the referenced code, controller, or workflow no longer exists in the codebase and no successor can be found, recommend Archive — don't default to Keep just because the general advice is still "sound." A learning about a deleted feature misleads readers into thinking that feature still exists. When in doubt between Keep and Archive, ask the user (in interactive mode) or mark as stale (in autonomous mode). But missing referenced files with no matching code is **not** a doubt case — it is strong, unambiguous Archive evidence. Auto-archive it.
|
||||
|
||||
## Scope Selection
|
||||
|
||||
Start by discovering learnings and pattern docs under `docs/solutions/`.
|
||||
|
||||
Exclude:
|
||||
|
||||
- `README.md`
|
||||
- `docs/solutions/_archived/`
|
||||
|
||||
Find all `.md` files under `docs/solutions/`, excluding `README.md` files and anything under `_archived/`.
|
||||
|
||||
If `$ARGUMENTS` is provided, use it to narrow scope before proceeding. Try these matching strategies in order, stopping at the first that produces results:
|
||||
|
||||
1. **Directory match** — check if the argument matches a subdirectory name under `docs/solutions/` (e.g., `performance-issues`, `database-issues`)
|
||||
2. **Frontmatter match** — search `module`, `component`, or `tags` fields in learning frontmatter for the argument
|
||||
3. **Filename match** — match against filenames (partial matches are fine)
|
||||
4. **Content search** — search file contents for the argument as a keyword (useful for feature names or feature areas)
|
||||
|
||||
If no matches are found, report that and ask the user to clarify. In autonomous mode, report the miss and stop — do not guess at scope.
|
||||
|
||||
If no candidate docs are found, report:
|
||||
|
||||
```text
|
||||
No candidate docs found in docs/solutions/.
|
||||
Run `ce:compound` after solving problems to start building your knowledge base.
|
||||
```
|
||||
|
||||
## Phase 0: Assess and Route
|
||||
|
||||
Before asking the user to classify anything:
|
||||
|
||||
1. Discover candidate artifacts
|
||||
2. Estimate scope
|
||||
3. Choose the lightest interaction path that fits
|
||||
|
||||
### Route by Scope
|
||||
|
||||
| Scope | When to use it | Interaction style |
|
||||
|-------|----------------|-------------------|
|
||||
| **Focused** | 1-2 likely files or user named a specific doc | Investigate directly, then present a recommendation |
|
||||
| **Batch** | Up to ~8 mostly independent docs | Investigate first, then present grouped recommendations |
|
||||
| **Broad** | 9+ docs, ambiguous, or repo-wide stale-doc sweep | Triage first, then investigate in batches |
|
||||
|
||||
### Broad Scope Triage
|
||||
|
||||
When scope is broad (9+ candidate docs), do a lightweight triage before deep investigation:
|
||||
|
||||
1. **Inventory** — read frontmatter of all candidate docs, group by module/component/category
|
||||
2. **Impact clustering** — identify areas with the densest clusters of learnings + pattern docs. A cluster of 5 learnings and 2 patterns covering the same module is higher-impact than 5 isolated single-doc areas, because staleness in one doc is likely to affect the others.
|
||||
3. **Spot-check drift** — for each cluster, check whether the primary referenced files still exist. Missing references in a high-impact cluster = strongest signal for where to start.
|
||||
4. **Recommend a starting area** — present the highest-impact cluster with a brief rationale and ask the user to confirm or redirect. In autonomous mode, skip the question and process all clusters in impact order.
|
||||
|
||||
Example:
|
||||
|
||||
```text
|
||||
Found 24 learnings across 5 areas.
|
||||
|
||||
The auth module has 5 learnings and 2 pattern docs that cross-reference
|
||||
each other — and 3 of those reference files that no longer exist.
|
||||
I'd start there.
|
||||
|
||||
1. Start with auth (recommended)
|
||||
2. Pick a different area
|
||||
3. Review everything
|
||||
```
|
||||
|
||||
Do not ask action-selection questions yet. First gather evidence.
|
||||
|
||||
## Phase 1: Investigate Candidate Learnings
|
||||
|
||||
For each learning in scope, read it, cross-reference its claims against the current codebase, and form a recommendation.
|
||||
|
||||
A learning has several dimensions that can independently go stale. Surface-level checks catch the obvious drift, but staleness often hides deeper:
|
||||
|
||||
- **References** — do the file paths, class names, and modules it mentions still exist or have they moved?
|
||||
- **Recommended solution** — does the fix still match how the code actually works today? A renamed file with a completely different implementation pattern is not just a path update.
|
||||
- **Code examples** — if the learning includes code snippets, do they still reflect the current implementation?
|
||||
- **Related docs** — are cross-referenced learnings and patterns still present and consistent?
|
||||
|
||||
Match investigation depth to the learning's specificity — a learning referencing exact file paths and code snippets needs more verification than one describing a general principle.
|
||||
|
||||
### Drift Classification: Update vs Replace
|
||||
|
||||
The critical distinction is whether the drift is **cosmetic** (references moved but the solution is the same) or **substantive** (the solution itself changed):
|
||||
|
||||
- **Update territory** — file paths moved, classes renamed, links broke, metadata drifted, but the core recommended approach is still how the code works. `ce:compound-refresh` fixes these directly.
|
||||
- **Replace territory** — the recommended solution conflicts with current code, the architectural approach changed, or the pattern is no longer the preferred way. This means a new learning needs to be written. A replacement subagent writes the successor following `ce:compound`'s document format (frontmatter, problem, root cause, solution, prevention), using the investigation evidence already gathered. The orchestrator does not rewrite learnings inline — it delegates to a subagent for context isolation.
|
||||
|
||||
**The boundary:** if you find yourself rewriting the solution section or changing what the learning recommends, stop — that is Replace, not Update.
|
||||
|
||||
### Judgment Guidelines
|
||||
|
||||
Three guidelines that are easy to get wrong:
|
||||
|
||||
1. **Contradiction = strong Replace signal.** If the learning's recommendation conflicts with current code patterns or a recently verified fix, that is not a minor drift — the learning is actively misleading. Classify as Replace.
|
||||
2. **Age alone is not a stale signal.** A 2-year-old learning that still matches current code is fine. Only use age as a prompt to inspect more carefully.
|
||||
3. **Check for successors before archiving.** Before recommending Replace or Archive, look for newer learnings, pattern docs, PRs, or issues covering the same problem space. If successor evidence exists, prefer Replace over Archive so readers are directed to the newer guidance.
|
||||
|
||||
## Phase 1.5: Investigate Pattern Docs
|
||||
|
||||
After reviewing the underlying learning docs, investigate any relevant pattern docs under `docs/solutions/patterns/`.
|
||||
|
||||
Pattern docs are high-leverage — a stale pattern is more dangerous than a stale individual learning because future work may treat it as broadly applicable guidance. Evaluate whether the generalized rule still holds given the refreshed state of the learnings it depends on.
|
||||
|
||||
A pattern doc with no clear supporting learnings is a stale signal — investigate carefully before keeping it unchanged.
|
||||
|
||||
## Subagent Strategy
|
||||
|
||||
Use subagents for context isolation when investigating multiple artifacts — not just because the task sounds complex. Choose the lightest approach that fits:
|
||||
|
||||
| Approach | When to use |
|
||||
|----------|-------------|
|
||||
| **Main thread only** | Small scope, short docs |
|
||||
| **Sequential subagents** | 1-2 artifacts with many supporting files to read |
|
||||
| **Parallel subagents** | 3+ truly independent artifacts with low overlap |
|
||||
| **Batched subagents** | Broad sweeps — narrow scope first, then investigate in batches |
|
||||
|
||||
**When spawning any subagent, include this instruction in its task prompt:**
|
||||
|
||||
> Use dedicated file search and read tools (Glob, Grep, Read) for all investigation. Do NOT use shell commands (ls, find, cat, grep, test, bash) for file operations. This avoids permission prompts and is more reliable.
|
||||
|
||||
There are two subagent roles:
|
||||
|
||||
1. **Investigation subagents** — read-only. They must not edit files, create successors, or archive anything. Each returns: file path, evidence, recommended action, confidence, and open questions. These can run in parallel when artifacts are independent.
|
||||
2. **Replacement subagents** — write a single new learning to replace a stale one. These run **one at a time, sequentially** (each replacement subagent may need to read significant code, and running multiple in parallel risks context exhaustion). The orchestrator handles all archival and metadata updates after each replacement completes.
|
||||
|
||||
The orchestrator merges investigation results, detects contradictions, coordinates replacement subagents, and performs all archival/metadata edits centrally. In interactive mode, it asks the user questions on ambiguous cases. In autonomous mode, it marks ambiguous cases as stale instead. If two artifacts overlap or discuss the same root issue, investigate them together rather than parallelizing.
|
||||
|
||||
## Phase 2: Classify the Right Maintenance Action
|
||||
|
||||
After gathering evidence, assign one recommended action.
|
||||
|
||||
### Keep
|
||||
|
||||
The learning is still accurate and useful. Do not edit the file — report that it was reviewed and remains trustworthy. Only add `last_refreshed` if you are already making a meaningful update for another reason.
|
||||
|
||||
### Update
|
||||
|
||||
The core solution is still valid but references have drifted (paths, class names, links, code snippets, metadata). Apply the fixes directly.
|
||||
|
||||
### Replace
|
||||
|
||||
Choose **Replace** when the learning's core guidance is now misleading — the recommended fix changed materially, the root cause or architecture shifted, or the preferred pattern is different.
|
||||
|
||||
The user may have invoked the refresh months after the original learning was written. Do not ask them for replacement context they are unlikely to have — use agent intelligence to investigate the codebase and synthesize the replacement.
|
||||
|
||||
**Evidence assessment:**
|
||||
|
||||
By the time you identify a Replace candidate, Phase 1 investigation has already gathered significant evidence: the old learning's claims, what the current code actually does, and where the drift occurred. Assess whether this evidence is sufficient to write a trustworthy replacement:
|
||||
|
||||
- **Sufficient evidence** — you understand both what the old learning recommended AND what the current approach is. The investigation found the current code patterns, the new file locations, the changed architecture. → Proceed to write the replacement (see Phase 4 Replace Flow).
|
||||
- **Insufficient evidence** — the drift is so fundamental that you cannot confidently document the current approach. The entire subsystem was replaced, or the new architecture is too complex to understand from a file scan alone. → Mark as stale in place:
|
||||
- Add `status: stale`, `stale_reason: [what you found]`, `stale_date: YYYY-MM-DD` to the frontmatter
|
||||
- Report what evidence you found and what is missing
|
||||
- Recommend the user run `ce:compound` after their next encounter with that area, when they have fresh problem-solving context
|
||||
|
||||
### Archive
|
||||
|
||||
Choose **Archive** when:
|
||||
|
||||
- The code or workflow no longer exists
|
||||
- The learning is obsolete and has no modern replacement worth documenting
|
||||
- The learning is redundant and no longer useful on its own
|
||||
- There is no meaningful successor evidence suggesting it should be replaced instead
|
||||
|
||||
Action:
|
||||
|
||||
- Move the file to `docs/solutions/_archived/`, preserving directory structure when helpful
|
||||
- Add:
|
||||
- `archived_date: YYYY-MM-DD`
|
||||
- `archive_reason: [why it was archived]`
|
||||
|
||||
### Before archiving: check if the problem domain is still active
|
||||
|
||||
When a learning's referenced files are gone, that is strong evidence — but only that the **implementation** is gone. Before archiving, reason about whether the **problem the learning solves** is still a concern in the codebase:
|
||||
|
||||
- A learning about session token storage where `auth_token.rb` is gone — does the application still handle session tokens? If so, the concept persists under a new implementation. That is Replace, not Archive.
|
||||
- A learning about a deprecated API endpoint where the entire feature was removed — the problem domain is gone. That is Archive.
|
||||
|
||||
Do not search mechanically for keywords from the old learning. Instead, understand what problem the learning addresses, then investigate whether that problem domain still exists in the codebase. The agent understands concepts — use that understanding to look for where the problem lives now, not where the old code used to be.
|
||||
|
||||
**Auto-archive only when both the implementation AND the problem domain are gone:**
|
||||
|
||||
- the referenced code is gone AND the application no longer deals with that problem domain
|
||||
- the learning is fully superseded by a clearly better successor
|
||||
- the document is plainly redundant and adds no distinct value
|
||||
|
||||
If the implementation is gone but the problem domain persists (the app still does auth, still processes payments, still handles migrations), classify as **Replace** — the problem still matters and the current approach should be documented.
|
||||
|
||||
Do not keep a learning just because its general advice is "still sound" — if the specific code it references is gone, the learning misleads readers. But do not archive a learning whose problem domain is still active — that knowledge gap should be filled with a replacement.
|
||||
|
||||
If there is a clearly better successor, strongly consider **Replace** before **Archive** so the old artifact points readers toward the newer guidance.
|
||||
|
||||
## Pattern Guidance
|
||||
|
||||
Apply the same four outcomes (Keep, Update, Replace, Archive) to pattern docs, but evaluate them as **derived guidance** rather than incident-level learnings. Key differences:
|
||||
|
||||
- **Keep**: the underlying learnings still support the generalized rule and examples remain representative
|
||||
- **Update**: the rule holds but examples, links, scope, or supporting references drifted
|
||||
- **Replace**: the generalized rule is now misleading, or the underlying learnings support a different synthesis. Base the replacement on the refreshed learning set — do not invent new rules from guesswork
|
||||
- **Archive**: the pattern is no longer valid, no longer recurring, or fully subsumed by a stronger pattern doc
|
||||
|
||||
If "archive" feels too strong but the pattern should no longer be elevated, reduce its prominence in place if the docs structure supports that.
|
||||
|
||||
## Phase 3: Ask for Decisions
|
||||
|
||||
### Autonomous mode
|
||||
|
||||
**Skip this entire phase. Do not ask any questions. Do not present options. Do not wait for input.** Proceed directly to Phase 4 and execute all actions based on the classifications from Phase 2:
|
||||
|
||||
- Unambiguous Keep, Update, auto-Archive, and Replace (with sufficient evidence) → execute directly
|
||||
- Ambiguous cases → mark as stale
|
||||
- Then generate the report (see Output Format)
|
||||
|
||||
### Interactive mode
|
||||
|
||||
Most Updates should be applied directly without asking. Only ask the user when:
|
||||
|
||||
- The right action is genuinely ambiguous (Update vs Replace vs Archive)
|
||||
- You are about to Archive a document **and** the evidence is not unambiguous (see auto-archive criteria in Phase 2). When auto-archive criteria are met, proceed without asking.
|
||||
- You are about to create a successor via `ce:compound`
|
||||
|
||||
Do **not** ask questions about whether code changes were intentional, whether the user wants to fix bugs in the code, or other concerns outside doc maintenance. Stay in your lane — doc accuracy.
|
||||
|
||||
#### Question Style
|
||||
|
||||
Always present choices using the platform's blocking question tool when available (`AskUserQuestion` in Claude Code, `request_user_input` in Codex, `ask_user` in Gemini). Otherwise, present numbered options in plain text and wait for the user's reply before proceeding.
|
||||
|
||||
Question rules:
|
||||
|
||||
- Ask **one question at a time**
|
||||
- Prefer **multiple choice**
|
||||
- Lead with the **recommended option**
|
||||
- Explain the rationale for the recommendation in one concise sentence
|
||||
- Avoid asking the user to choose from actions that are not actually plausible
|
||||
|
||||
#### Focused Scope
|
||||
|
||||
For a single artifact, present:
|
||||
|
||||
- file path
|
||||
- 2-4 bullets of evidence
|
||||
- recommended action
|
||||
|
||||
Then ask:
|
||||
|
||||
```text
|
||||
This [learning/pattern] looks like a [Update/Keep/Replace/Archive].
|
||||
|
||||
Why: [one-sentence rationale based on the evidence]
|
||||
|
||||
What would you like to do?
|
||||
|
||||
1. [Recommended action]
|
||||
2. [Second plausible action]
|
||||
3. Skip for now
|
||||
```
|
||||
|
||||
Do not list all four actions unless all four are genuinely plausible.
|
||||
|
||||
#### Batch Scope
|
||||
|
||||
For several learnings:
|
||||
|
||||
1. Group obvious **Keep** cases together
|
||||
2. Group obvious **Update** cases together when the fixes are straightforward
|
||||
3. Present **Replace** cases individually or in very small groups
|
||||
4. Present **Archive** cases individually unless they are strong auto-archive candidates
|
||||
|
||||
Ask for confirmation in stages:
|
||||
|
||||
1. Confirm grouped Keep/Update recommendations
|
||||
2. Then handle Replace one at a time
|
||||
3. Then handle Archive one at a time unless the archive is unambiguous and safe to auto-apply
|
||||
|
||||
#### Broad Scope
|
||||
|
||||
If the user asked for a sweeping refresh, keep the interaction incremental:
|
||||
|
||||
1. Narrow scope first
|
||||
2. Investigate a manageable batch
|
||||
3. Present recommendations
|
||||
4. Ask whether to continue to the next batch
|
||||
|
||||
Do not front-load the user with a full maintenance queue.
|
||||
|
||||
## Phase 4: Execute the Chosen Action
|
||||
|
||||
### Keep Flow
|
||||
|
||||
No file edit by default. Summarize why the learning remains trustworthy.
|
||||
|
||||
### Update Flow
|
||||
|
||||
Apply in-place edits only when the solution is still substantively correct.
|
||||
|
||||
Examples of valid in-place updates:
|
||||
|
||||
- Rename `app/models/auth_token.rb` reference to `app/models/session_token.rb`
|
||||
- Update `module: AuthToken` to `module: SessionToken`
|
||||
- Fix outdated links to related docs
|
||||
- Refresh implementation notes after a directory move
|
||||
|
||||
Examples that should **not** be in-place updates:
|
||||
|
||||
- Fixing a typo with no effect on understanding
|
||||
- Rewording prose for style alone
|
||||
- Small cleanup that does not materially improve accuracy or usability
|
||||
- The old fix is now an anti-pattern
|
||||
- The system architecture changed enough that the old guidance is misleading
|
||||
- The troubleshooting path is materially different
|
||||
|
||||
Those cases require **Replace**, not Update.
|
||||
|
||||
### Replace Flow
|
||||
|
||||
Process Replace candidates **one at a time, sequentially**. Each replacement is written by a subagent to protect the main context window.
|
||||
|
||||
**When evidence is sufficient:**
|
||||
|
||||
1. Spawn a single subagent to write the replacement learning. Pass it:
|
||||
- The old learning's full content
|
||||
- A summary of the investigation evidence (what changed, what the current code does, why the old guidance is misleading)
|
||||
- The target path and category (same category as the old learning unless the category itself changed)
|
||||
2. The subagent writes the new learning following `ce:compound`'s document format: YAML frontmatter (title, category, date, module, component, tags), problem description, root cause, current solution with code examples, and prevention tips. It should use dedicated file search and read tools if it needs additional context beyond what was passed.
|
||||
3. After the subagent completes, the orchestrator:
|
||||
- Adds `superseded_by: [new learning path]` to the old learning's frontmatter
|
||||
- Moves the old learning to `docs/solutions/_archived/`
|
||||
|
||||
**When evidence is insufficient:**
|
||||
|
||||
1. Mark the learning as stale in place:
|
||||
- Add to frontmatter: `status: stale`, `stale_reason: [what you found]`, `stale_date: YYYY-MM-DD`
|
||||
2. Report what evidence was found and what is missing
|
||||
3. Recommend the user run `ce:compound` after their next encounter with that area
|
||||
|
||||
### Archive Flow
|
||||
|
||||
Archive only when a learning is clearly obsolete or redundant. Do not archive a document just because it is old.
|
||||
|
||||
## Output Format
|
||||
|
||||
**The full report MUST be printed as markdown output.** Do not summarize findings internally and then output a one-liner. The report is the deliverable — print every section in full, formatted as readable markdown with headers, tables, and bullet points.
|
||||
|
||||
After processing the selected scope, output the following report:
|
||||
|
||||
```text
|
||||
Compound Refresh Summary
|
||||
========================
|
||||
Scanned: N learnings
|
||||
|
||||
Kept: X
|
||||
Updated: Y
|
||||
Replaced: Z
|
||||
Archived: W
|
||||
Skipped: V
|
||||
Marked stale: S
|
||||
```
|
||||
|
||||
Then for EVERY file processed, list:
|
||||
- The file path
|
||||
- The classification (Keep/Update/Replace/Archive/Stale)
|
||||
- What evidence was found
|
||||
- What action was taken (or recommended)
|
||||
|
||||
For **Keep** outcomes, list them under a reviewed-without-edits section so the result is visible without creating git churn.
|
||||
|
||||
### Autonomous mode output
|
||||
|
||||
In autonomous mode, the report is the sole deliverable — there is no user present to ask follow-up questions, so the report must be self-contained and complete. **Print the full report. Do not abbreviate, summarize, or skip sections.**
|
||||
|
||||
Split actions into two sections:
|
||||
|
||||
**Applied** (writes that succeeded):
|
||||
- For each **Updated** file: the file path, what references were fixed, and why
|
||||
- For each **Replaced** file: what the old learning recommended vs what the current code does, and the path to the new successor
|
||||
- For each **Archived** file: the file path and what referenced code/workflow is gone
|
||||
- For each **Marked stale** file: the file path, what evidence was found, and why it was ambiguous
|
||||
|
||||
**Recommended** (actions that could not be written — e.g., permission denied):
|
||||
- Same detail as above, but framed as recommendations for a human to apply
|
||||
- Include enough context that the user can apply the change manually or re-run the skill interactively
|
||||
|
||||
If all writes succeed, the Recommended section is empty. If no writes succeed (e.g., read-only invocation), all actions appear under Recommended — the report becomes a maintenance plan.
|
||||
|
||||
## Phase 5: Commit Changes
|
||||
|
||||
After all actions are executed and the report is generated, handle committing the changes. Skip this phase if no files were modified (all Keep, or all writes failed).
|
||||
|
||||
### Detect git context
|
||||
|
||||
Before offering options, check:
|
||||
1. Which branch is currently checked out (main/master vs feature branch)
|
||||
2. Whether the working tree has other uncommitted changes beyond what compound-refresh modified
|
||||
3. Recent commit messages to match the repo's commit style
|
||||
|
||||
### Autonomous mode
|
||||
|
||||
Use sensible defaults — no user to ask:
|
||||
|
||||
| Context | Default action |
|
||||
|---------|---------------|
|
||||
| On main/master | Create a branch named for what was refreshed (e.g., `docs/refresh-auth-and-ci-learnings`), commit, attempt to open a PR. If PR creation fails, report the branch name. |
|
||||
| On a feature branch | Commit as a separate commit on the current branch |
|
||||
| Git operations fail | Include the recommended git commands in the report and continue |
|
||||
|
||||
Stage only the files that compound-refresh modified — not other dirty files in the working tree.
|
||||
|
||||
### Interactive mode
|
||||
|
||||
First, run `git branch --show-current` to determine the current branch. Then present the correct options based on the result. Stage only compound-refresh files regardless of which option the user picks.
|
||||
|
||||
**If the current branch is main, master, or the repo's default branch:**
|
||||
|
||||
1. Create a branch, commit, and open a PR (recommended) — the branch name should be specific to what was refreshed, not generic (e.g., `docs/refresh-auth-learnings` not `docs/compound-refresh`)
|
||||
2. Commit directly to `{current branch name}`
|
||||
3. Don't commit — I'll handle it
|
||||
|
||||
**If the current branch is a feature branch, clean working tree:**
|
||||
|
||||
1. Commit to `{current branch name}` as a separate commit (recommended)
|
||||
2. Create a separate branch and commit
|
||||
3. Don't commit
|
||||
|
||||
**If the current branch is a feature branch, dirty working tree (other uncommitted changes):**
|
||||
|
||||
1. Commit only the compound-refresh changes to `{current branch name}` (selective staging — other dirty files stay untouched)
|
||||
2. Don't commit
|
||||
|
||||
### Commit message
|
||||
|
||||
Write a descriptive commit message that:
|
||||
- Summarizes what was refreshed (e.g., "update 3 stale learnings, archive 1 obsolete doc")
|
||||
- Follows the repo's existing commit conventions (check recent git log for style)
|
||||
- Is succinct — the details are in the changed files themselves
|
||||
|
||||
## Relationship to ce:compound
|
||||
|
||||
- `ce:compound` captures a newly solved, verified problem
|
||||
- `ce:compound-refresh` maintains older learnings as the codebase evolves
|
||||
|
||||
Use **Replace** only when the refresh process has enough real evidence to write a trustworthy successor. When evidence is insufficient, mark as stale and recommend `ce:compound` for when the user next encounters that problem area.
|
||||
@@ -1,5 +1,5 @@
|
||||
---
|
||||
name: workflows:compound
|
||||
name: ce:compound
|
||||
description: Document a recently solved problem to compound your team's knowledge
|
||||
argument-hint: "[optional: brief context about the fix]"
|
||||
---
|
||||
@@ -17,11 +17,19 @@ Captures problem solutions while context is fresh, creating structured documenta
|
||||
## Usage
|
||||
|
||||
```bash
|
||||
/workflows:compound # Document the most recent fix
|
||||
/workflows:compound [brief context] # Provide additional context hint
|
||||
/ce:compound # Document the most recent fix
|
||||
/ce:compound [brief context] # Provide additional context hint
|
||||
```
|
||||
|
||||
## Execution Strategy: Two-Phase Orchestration
|
||||
## Execution Strategy
|
||||
|
||||
**Always run full mode by default.** Proceed directly to Phase 1 unless the user explicitly requests compact-safe mode (e.g., `/ce:compound --compact` or "use compact mode").
|
||||
|
||||
Compact-safe mode exists as a lightweight alternative — see the **Compact-Safe Mode** section below. It's there if the user wants it, not something to push.
|
||||
|
||||
---
|
||||
|
||||
### Full Mode
|
||||
|
||||
<critical_requirement>
|
||||
**Only ONE file gets written - the final documentation.**
|
||||
@@ -51,7 +59,8 @@ Launch these subagents IN PARALLEL. Each returns text data to the orchestrator.
|
||||
- Searches `docs/solutions/` for related documentation
|
||||
- Identifies cross-references and links
|
||||
- Finds related GitHub issues
|
||||
- Returns: Links and relationships
|
||||
- Flags any related learning or pattern docs that may now be stale, contradicted, or overly broad
|
||||
- Returns: Links, relationships, and any refresh candidates
|
||||
|
||||
#### 4. **Prevention Strategist**
|
||||
- Develops prevention strategies
|
||||
@@ -83,6 +92,53 @@ The orchestrating agent (main conversation) performs these steps:
|
||||
|
||||
</sequential_tasks>
|
||||
|
||||
### Phase 2.5: Selective Refresh Check
|
||||
|
||||
After writing the new learning, decide whether this new solution is evidence that older docs should be refreshed.
|
||||
|
||||
`ce:compound-refresh` is **not** a default follow-up. Use it selectively when the new learning suggests an older learning or pattern doc may now be inaccurate.
|
||||
|
||||
It makes sense to invoke `ce:compound-refresh` when one or more of these are true:
|
||||
|
||||
1. A related learning or pattern doc recommends an approach that the new fix now contradicts
|
||||
2. The new fix clearly supersedes an older documented solution
|
||||
3. The current work involved a refactor, migration, rename, or dependency upgrade that likely invalidated references in older docs
|
||||
4. A pattern doc now looks overly broad, outdated, or no longer supported by the refreshed reality
|
||||
5. The Related Docs Finder surfaced high-confidence refresh candidates in the same problem space
|
||||
|
||||
It does **not** make sense to invoke `ce:compound-refresh` when:
|
||||
|
||||
1. No related docs were found
|
||||
2. Related docs still appear consistent with the new learning
|
||||
3. The overlap is superficial and does not change prior guidance
|
||||
4. Refresh would require a broad historical review with weak evidence
|
||||
|
||||
Use these rules:
|
||||
|
||||
- If there is **one obvious stale candidate**, invoke `ce:compound-refresh` with a narrow scope hint after the new learning is written
|
||||
- If there are **multiple candidates in the same area**, ask the user whether to run a targeted refresh for that module, category, or pattern set
|
||||
- If context is already tight or you are in compact-safe mode, do not expand into a broad refresh automatically; instead recommend `ce:compound-refresh` as the next step with a scope hint
|
||||
|
||||
When invoking or recommending `ce:compound-refresh`, be explicit about the argument to pass. Prefer the narrowest useful scope:
|
||||
|
||||
- **Specific file** when one learning or pattern doc is the likely stale artifact
|
||||
- **Module or component name** when several related docs may need review
|
||||
- **Category name** when the drift is concentrated in one solutions area
|
||||
- **Pattern filename or pattern topic** when the stale guidance lives in `docs/solutions/patterns/`
|
||||
|
||||
Examples:
|
||||
|
||||
- `/ce:compound-refresh plugin-versioning-requirements`
|
||||
- `/ce:compound-refresh payments`
|
||||
- `/ce:compound-refresh performance-issues`
|
||||
- `/ce:compound-refresh critical-patterns`
|
||||
|
||||
A single scope hint may still expand to multiple related docs when the change is cross-cutting within one domain, category, or pattern area.
|
||||
|
||||
Do not invoke `ce:compound-refresh` without an argument unless the user explicitly wants a broad sweep.
|
||||
|
||||
Always capture the new learning first. Refresh is a targeted maintenance follow-up, not a prerequisite for documentation.
|
||||
|
||||
### Phase 3: Optional Enhancement
|
||||
|
||||
**WAIT for Phase 2 to complete before proceeding.**
|
||||
@@ -99,6 +155,46 @@ Based on problem type, optionally invoke specialized agents to review the docume
|
||||
|
||||
</parallel_tasks>
|
||||
|
||||
---
|
||||
|
||||
### Compact-Safe Mode
|
||||
|
||||
<critical_requirement>
|
||||
**Single-pass alternative for context-constrained sessions.**
|
||||
|
||||
When context budget is tight, this mode skips parallel subagents entirely. The orchestrator performs all work in a single pass, producing a minimal but complete solution document.
|
||||
</critical_requirement>
|
||||
|
||||
The orchestrator (main conversation) performs ALL of the following in one sequential pass:
|
||||
|
||||
1. **Extract from conversation**: Identify the problem, root cause, and solution from conversation history
|
||||
2. **Classify**: Determine category and filename (same categories as full mode)
|
||||
3. **Write minimal doc**: Create `docs/solutions/[category]/[filename].md` with:
|
||||
- YAML frontmatter (title, category, date, tags)
|
||||
- Problem description (1-2 sentences)
|
||||
- Root cause (1-2 sentences)
|
||||
- Solution with key code snippets
|
||||
- One prevention tip
|
||||
4. **Skip specialized agent reviews** (Phase 3) to conserve context
|
||||
|
||||
**Compact-safe output:**
|
||||
```
|
||||
✓ Documentation complete (compact-safe mode)
|
||||
|
||||
File created:
|
||||
- docs/solutions/[category]/[filename].md
|
||||
|
||||
Note: This was created in compact-safe mode. For richer documentation
|
||||
(cross-references, detailed prevention strategies, specialized reviews),
|
||||
re-run /compound in a fresh session.
|
||||
```
|
||||
|
||||
**No subagents are launched. No parallel tasks. One file written.**
|
||||
|
||||
In compact-safe mode, only suggest `ce:compound-refresh` if there is an obvious narrow refresh target. Do not broaden into a large refresh sweep from a compact-safe session.
|
||||
|
||||
---
|
||||
|
||||
## What It Captures
|
||||
|
||||
- **Problem symptom**: Exact error messages, observable behavior
|
||||
@@ -203,7 +299,7 @@ Build → Test → Find Issue → Research → Improve → Document → Validate
|
||||
|
||||
<auto_invoke> <trigger_phrases> - "that worked" - "it's fixed" - "working now" - "problem solved" </trigger_phrases>
|
||||
|
||||
<manual_override> Use /workflows:compound [context] to document immediately without waiting for auto-detection. </manual_override> </auto_invoke>
|
||||
<manual_override> Use /ce:compound [context] to document immediately without waiting for auto-detection. </manual_override> </auto_invoke>
|
||||
|
||||
## Routes To
|
||||
|
||||
@@ -231,10 +327,10 @@ Based on problem type, these agents can enhance documentation:
|
||||
|
||||
### When to Invoke
|
||||
- **Auto-triggered** (optional): Agents can run post-documentation for enhancement
|
||||
- **Manual trigger**: User can invoke agents after /workflows:compound completes for deeper review
|
||||
- **Manual trigger**: User can invoke agents after /ce:compound completes for deeper review
|
||||
- **Customize agents**: Edit `compound-engineering.local.md` or invoke the `setup` skill to configure which review agents are used across all workflows
|
||||
|
||||
## Related Commands
|
||||
|
||||
- `/research [topic]` - Deep investigation (searches docs/solutions/ for patterns)
|
||||
- `/workflows:plan` - Planning workflow (references documented solutions)
|
||||
- `/ce:plan` - Planning workflow (references documented solutions)
|
||||
370
plugins/compound-engineering/skills/ce-ideate/SKILL.md
Normal file
370
plugins/compound-engineering/skills/ce-ideate/SKILL.md
Normal file
@@ -0,0 +1,370 @@
|
||||
---
|
||||
name: ce:ideate
|
||||
description: "Generate and critically evaluate grounded improvement ideas for the current project. Use when asking what to improve, requesting idea generation, exploring surprising improvements, or wanting the AI to proactively suggest strong project directions before brainstorming one in depth. Triggers on phrases like 'what should I improve', 'give me ideas', 'ideate on this project', 'surprise me with improvements', 'what would you change', or any request for AI-generated project improvement suggestions rather than refining the user's own idea."
|
||||
argument-hint: "[optional: feature, focus area, or constraint]"
|
||||
---
|
||||
|
||||
# Generate Improvement Ideas
|
||||
|
||||
**Note: The current year is 2026.** Use this when dating ideation documents and checking recent ideation artifacts.
|
||||
|
||||
`ce:ideate` precedes `ce:brainstorm`.
|
||||
|
||||
- `ce:ideate` answers: "What are the strongest ideas worth exploring?"
|
||||
- `ce:brainstorm` answers: "What exactly should one chosen idea mean?"
|
||||
- `ce:plan` answers: "How should it be built?"
|
||||
|
||||
This workflow produces a ranked ideation artifact in `docs/ideation/`. It does **not** produce requirements, plans, or code.
|
||||
|
||||
## Interaction Method
|
||||
|
||||
Use the platform's blocking question tool when available (`AskUserQuestion` in Claude Code, `request_user_input` in Codex, `ask_user` in Gemini). Otherwise, present numbered options in chat and wait for the user's reply before proceeding.
|
||||
|
||||
Ask one question at a time. Prefer concise single-select choices when natural options exist.
|
||||
|
||||
## Focus Hint
|
||||
|
||||
<focus_hint> #$ARGUMENTS </focus_hint>
|
||||
|
||||
Interpret any provided argument as optional context. It may be:
|
||||
|
||||
- a concept such as `DX improvements`
|
||||
- a path such as `plugins/compound-engineering/skills/`
|
||||
- a constraint such as `low-complexity quick wins`
|
||||
- a volume hint such as `top 3`, `100 ideas`, or `raise the bar`
|
||||
|
||||
If no argument is provided, proceed with open-ended ideation.
|
||||
|
||||
## Core Principles
|
||||
|
||||
1. **Ground before ideating** - Scan the actual codebase first. Do not generate abstract product advice detached from the repository.
|
||||
2. **Diverge before judging** - Generate the full idea set before evaluating any individual idea.
|
||||
3. **Use adversarial filtering** - The quality mechanism is explicit rejection with reasons, not optimistic ranking.
|
||||
4. **Preserve the original prompt mechanism** - Generate many ideas, critique the whole list, then explain only the survivors in detail. Do not let extra process obscure this pattern.
|
||||
5. **Use agent diversity to improve the candidate pool** - Parallel sub-agents are a support mechanism for richer idea generation and critique, not the core workflow itself.
|
||||
6. **Preserve the artifact early** - Write the ideation document before presenting results so work survives interruptions.
|
||||
7. **Route action into brainstorming** - Ideation identifies promising directions; `ce:brainstorm` defines the selected one precisely enough for planning.
|
||||
|
||||
## Execution Flow
|
||||
|
||||
### Phase 0: Resume and Scope
|
||||
|
||||
#### 0.1 Check for Recent Ideation Work
|
||||
|
||||
Look in `docs/ideation/` for ideation documents created within the last 30 days.
|
||||
|
||||
Treat a prior ideation doc as relevant when:
|
||||
- the topic matches the requested focus
|
||||
- the path or subsystem overlaps the requested focus
|
||||
- the request is open-ended and there is an obvious recent open ideation doc
|
||||
- the issue-grounded status matches: do not offer to resume a non-issue ideation when the current argument indicates issue-tracker intent, or vice versa — treat these as distinct topics
|
||||
|
||||
If a relevant doc exists, ask whether to:
|
||||
1. continue from it
|
||||
2. start fresh
|
||||
|
||||
If continuing:
|
||||
- read the document
|
||||
- summarize what has already been explored
|
||||
- preserve previous idea statuses and session log entries
|
||||
- update the existing file instead of creating a duplicate
|
||||
|
||||
#### 0.2 Interpret Focus and Volume
|
||||
|
||||
Infer three things from the argument:
|
||||
|
||||
- **Focus context** - concept, path, constraint, or open-ended
|
||||
- **Volume override** - any hint that changes candidate or survivor counts
|
||||
- **Issue-tracker intent** - whether the user wants issue/bug data as an input source
|
||||
|
||||
Issue-tracker intent triggers when the argument's primary intent is about analyzing issue patterns: `bugs`, `github issues`, `open issues`, `issue patterns`, `what users are reporting`, `bug reports`, `issue themes`.
|
||||
|
||||
Do NOT trigger on arguments that merely mention bugs as a focus: `bug in auth`, `fix the login issue`, `the signup bug` — these are focus hints, not requests to analyze the issue tracker.
|
||||
|
||||
When combined (e.g., `top 3 bugs in authentication`): detect issue-tracker intent first, volume override second, remainder is the focus hint. The focus narrows which issues matter; the volume override controls survivor count.
|
||||
|
||||
Default volume:
|
||||
- each ideation sub-agent generates about 7-8 ideas (yielding 30-40 raw ideas across agents, ~20-30 after dedupe)
|
||||
- keep the top 5-7 survivors
|
||||
|
||||
Honor clear overrides such as:
|
||||
- `top 3`
|
||||
- `100 ideas`
|
||||
- `go deep`
|
||||
- `raise the bar`
|
||||
|
||||
Use reasonable interpretation rather than formal parsing.
|
||||
|
||||
### Phase 1: Codebase Scan
|
||||
|
||||
Before generating ideas, gather codebase context.
|
||||
|
||||
Run agents in parallel in the **foreground** (do not use background dispatch — the results are needed before proceeding):
|
||||
|
||||
1. **Quick context scan** — dispatch a general-purpose sub-agent with this prompt:
|
||||
|
||||
> Read the project's AGENTS.md (or CLAUDE.md only as compatibility fallback, then README.md if neither exists), then discover the top-level directory layout using the native file-search/glob tool (e.g., `Glob` with pattern `*` or `*/*` in Claude Code). Return a concise summary (under 30 lines) covering:
|
||||
> - project shape (language, framework, top-level directory layout)
|
||||
> - notable patterns or conventions
|
||||
> - obvious pain points or gaps
|
||||
> - likely leverage points for improvement
|
||||
>
|
||||
> Keep the scan shallow — read only top-level documentation and directory structure. Do not analyze GitHub issues, templates, or contribution guidelines. Do not do deep code search.
|
||||
>
|
||||
> Focus hint: {focus_hint}
|
||||
|
||||
2. **Learnings search** — dispatch `compound-engineering:research:learnings-researcher` with a brief summary of the ideation focus.
|
||||
|
||||
3. **Issue intelligence** (conditional) — if issue-tracker intent was detected in Phase 0.2, dispatch `compound-engineering:research:issue-intelligence-analyst` with the focus hint. If a focus hint is present, pass it so the agent can weight its clustering toward that area. Run this in parallel with agents 1 and 2.
|
||||
|
||||
If the agent returns an error (gh not installed, no remote, auth failure), log a warning to the user ("Issue analysis unavailable: {reason}. Proceeding with standard ideation.") and continue with the existing two-agent grounding.
|
||||
|
||||
If the agent reports fewer than 5 total issues, note "Insufficient issue signal for theme analysis" and proceed with default ideation frames in Phase 2.
|
||||
|
||||
Consolidate all results into a short grounding summary. When issue intelligence is present, keep it as a distinct section so ideation sub-agents can distinguish between code-observed and user-reported signals:
|
||||
|
||||
- **Codebase context** — project shape, notable patterns, obvious pain points, likely leverage points
|
||||
- **Past learnings** — relevant institutional knowledge from docs/solutions/
|
||||
- **Issue intelligence** (when present) — theme summaries from the issue intelligence agent, preserving theme titles, descriptions, issue counts, and trend directions
|
||||
|
||||
Do **not** do external research in v1.
|
||||
|
||||
### Phase 2: Divergent Ideation
|
||||
|
||||
Follow this mechanism exactly:
|
||||
|
||||
1. Generate the full candidate list before critiquing any idea.
|
||||
2. Each sub-agent targets about 7-8 ideas by default. With 4-6 agents this yields 30-40 raw ideas, which merge and dedupe to roughly 20-30 unique candidates. Adjust the per-agent target when volume overrides apply (e.g., "100 ideas" raises it, "top 3" may lower the survivor count instead).
|
||||
3. Push past the safe obvious layer. Each agent's first few ideas tend to be obvious — push past them.
|
||||
4. Ground every idea in the Phase 1 scan.
|
||||
5. Use this prompting pattern as the backbone:
|
||||
- first generate many ideas
|
||||
- then challenge them systematically
|
||||
- then explain only the survivors in detail
|
||||
6. If the platform supports sub-agents, use them to improve diversity in the candidate pool rather than to replace the core mechanism.
|
||||
7. Give each ideation sub-agent the same:
|
||||
- grounding summary
|
||||
- focus hint
|
||||
- per-agent volume target (~7-8 ideas by default)
|
||||
- instruction to generate raw candidates only, not critique
|
||||
8. When using sub-agents, assign each one a different ideation frame as a **starting bias, not a constraint**. Prompt each agent to begin from its assigned perspective but follow any promising thread wherever it leads — cross-cutting ideas that span multiple frames are valuable, not out of scope.
|
||||
|
||||
**Frame selection depends on whether issue intelligence is active:**
|
||||
|
||||
**When issue-tracker intent is active and themes were returned:**
|
||||
- Each theme with `confidence: high` or `confidence: medium` becomes an ideation frame. The frame prompt uses the theme title and description as the starting bias.
|
||||
- If fewer than 4 cluster-derived frames, pad with default frames in this order: "leverage and compounding effects", "assumption-breaking or reframing", "inversion, removal, or automation of a painful step". These complement issue-grounded themes by pushing beyond the reported problems.
|
||||
- Cap at 6 total frames. If more than 6 themes qualify, use the top 6 by issue count; note remaining themes in the grounding summary as "minor themes" so sub-agents are still aware of them.
|
||||
|
||||
**When issue-tracker intent is NOT active (default):**
|
||||
- user or operator pain and friction
|
||||
- unmet need or missing capability
|
||||
- inversion, removal, or automation of a painful step
|
||||
- assumption-breaking or reframing
|
||||
- leverage and compounding effects
|
||||
- extreme cases, edge cases, or power-user pressure
|
||||
9. Ask each ideation sub-agent to return a standardized structure for each idea so the orchestrator can merge and reason over the outputs consistently. Prefer a compact JSON-like structure with:
|
||||
- title
|
||||
- summary
|
||||
- why_it_matters
|
||||
- evidence or grounding hooks
|
||||
- optional local signals such as boldness or focus_fit
|
||||
10. Merge and dedupe the sub-agent outputs into one master candidate list.
|
||||
11. **Synthesize cross-cutting combinations.** After deduping, scan the merged list for ideas from different frames that together suggest something stronger than either alone. If two or more ideas naturally combine into a higher-leverage proposal, add the combined idea to the list (expect 3-5 additions at most). This synthesis step belongs to the orchestrator because it requires seeing all ideas simultaneously.
|
||||
12. Spread ideas across multiple dimensions when justified:
|
||||
- workflow/DX
|
||||
- reliability
|
||||
- extensibility
|
||||
- missing capabilities
|
||||
- docs/knowledge compounding
|
||||
- quality and maintenance
|
||||
- leverage on future work
|
||||
13. If a focus was provided, pass it to every ideation sub-agent and weight the merged list toward it without excluding stronger adjacent ideas.
|
||||
|
||||
The mechanism to preserve is:
|
||||
- generate many ideas first
|
||||
- critique the full combined list second
|
||||
- explain only the survivors in detail
|
||||
|
||||
The sub-agent pattern to preserve is:
|
||||
- independent ideation with frames as starting biases first
|
||||
- orchestrator merge, dedupe, and cross-cutting synthesis second
|
||||
- critique only after the combined and synthesized list exists
|
||||
|
||||
### Phase 3: Adversarial Filtering
|
||||
|
||||
Review every generated idea critically.
|
||||
|
||||
Prefer a two-layer critique:
|
||||
1. Have one or more skeptical sub-agents attack the merged list from distinct angles.
|
||||
2. Have the orchestrator synthesize those critiques, apply the rubric consistently, score the survivors, and decide the final ranking.
|
||||
|
||||
Do not let critique agents generate replacement ideas in this phase unless explicitly refining.
|
||||
|
||||
Critique agents may provide local judgments, but final scoring authority belongs to the orchestrator so the ranking stays consistent across different frames and perspectives.
|
||||
|
||||
For each rejected idea, write a one-line reason.
|
||||
|
||||
Use rejection criteria such as:
|
||||
- too vague
|
||||
- not actionable
|
||||
- duplicates a stronger idea
|
||||
- not grounded in the current codebase
|
||||
- too expensive relative to likely value
|
||||
- already covered by existing workflows or docs
|
||||
- interesting but better handled as a brainstorm variant, not a product improvement
|
||||
|
||||
Use a consistent survivor rubric that weighs:
|
||||
- groundedness in the current repo
|
||||
- expected value
|
||||
- novelty
|
||||
- pragmatism
|
||||
- leverage on future work
|
||||
- implementation burden
|
||||
- overlap with stronger ideas
|
||||
|
||||
Target output:
|
||||
- keep 5-7 survivors by default
|
||||
- if too many survive, run a second stricter pass
|
||||
- if fewer than 5 survive, report that honestly rather than lowering the bar
|
||||
|
||||
### Phase 4: Present the Survivors
|
||||
|
||||
Present the surviving ideas to the user before writing the durable artifact.
|
||||
|
||||
This first presentation is a review checkpoint, not the final archived result.
|
||||
|
||||
Present only the surviving ideas in structured form:
|
||||
|
||||
- title
|
||||
- description
|
||||
- rationale
|
||||
- downsides
|
||||
- confidence score
|
||||
- estimated complexity
|
||||
|
||||
Then include a brief rejection summary so the user can see what was considered and cut.
|
||||
|
||||
Keep the presentation concise. The durable artifact holds the full record.
|
||||
|
||||
Allow brief follow-up questions and lightweight clarification before writing the artifact.
|
||||
|
||||
Do not write the ideation doc yet unless:
|
||||
- the user indicates the candidate set is good enough to preserve
|
||||
- the user asks to refine and continue in a way that should be recorded
|
||||
- the workflow is about to hand off to `ce:brainstorm`, Proof sharing, or session end
|
||||
|
||||
### Phase 5: Write the Ideation Artifact
|
||||
|
||||
Write the ideation artifact after the candidate set has been reviewed enough to preserve.
|
||||
|
||||
Always write or update the artifact before:
|
||||
- handing off to `ce:brainstorm`
|
||||
- sharing to Proof
|
||||
- ending the session
|
||||
|
||||
To write the artifact:
|
||||
|
||||
1. Ensure `docs/ideation/` exists
|
||||
2. Choose the file path:
|
||||
- `docs/ideation/YYYY-MM-DD-<topic>-ideation.md`
|
||||
- `docs/ideation/YYYY-MM-DD-open-ideation.md` when no focus exists
|
||||
3. Write or update the ideation document
|
||||
|
||||
Use this structure and omit clearly irrelevant fields only when necessary:
|
||||
|
||||
```markdown
|
||||
---
|
||||
date: YYYY-MM-DD
|
||||
topic: <kebab-case-topic>
|
||||
focus: <optional focus hint>
|
||||
---
|
||||
|
||||
# Ideation: <Title>
|
||||
|
||||
## Codebase Context
|
||||
[Grounding summary from Phase 1]
|
||||
|
||||
## Ranked Ideas
|
||||
|
||||
### 1. <Idea Title>
|
||||
**Description:** [Concrete explanation]
|
||||
**Rationale:** [Why this improves the project]
|
||||
**Downsides:** [Tradeoffs or costs]
|
||||
**Confidence:** [0-100%]
|
||||
**Complexity:** [Low / Medium / High]
|
||||
**Status:** [Unexplored / Explored]
|
||||
|
||||
## Rejection Summary
|
||||
|
||||
| # | Idea | Reason Rejected |
|
||||
|---|------|-----------------|
|
||||
| 1 | <Idea> | <Reason rejected> |
|
||||
|
||||
## Session Log
|
||||
- YYYY-MM-DD: Initial ideation — <candidate count> generated, <survivor count> survived
|
||||
```
|
||||
|
||||
If resuming:
|
||||
- update the existing file in place
|
||||
- append to the session log
|
||||
- preserve explored markers
|
||||
|
||||
### Phase 6: Refine or Hand Off
|
||||
|
||||
After presenting the results, ask what should happen next.
|
||||
|
||||
Offer these options:
|
||||
1. brainstorm a selected idea
|
||||
2. refine the ideation
|
||||
3. share to Proof
|
||||
4. end the session
|
||||
|
||||
#### 6.1 Brainstorm a Selected Idea
|
||||
|
||||
If the user selects an idea:
|
||||
- write or update the ideation doc first
|
||||
- mark that idea as `Explored`
|
||||
- note the brainstorm date in the session log
|
||||
- invoke `ce:brainstorm` with the selected idea as the seed
|
||||
|
||||
Do **not** skip brainstorming and go straight to planning from ideation output.
|
||||
|
||||
#### 6.2 Refine the Ideation
|
||||
|
||||
Route refinement by intent:
|
||||
|
||||
- `add more ideas` or `explore new angles` -> return to Phase 2
|
||||
- `re-evaluate` or `raise the bar` -> return to Phase 3
|
||||
- `dig deeper on idea #N` -> expand only that idea's analysis
|
||||
|
||||
After each refinement:
|
||||
- update the ideation document before any handoff, sharing, or session end
|
||||
- append a session log entry
|
||||
|
||||
#### 6.3 Share to Proof
|
||||
|
||||
If requested, share the ideation document using the standard Proof markdown upload pattern already used elsewhere in the plugin.
|
||||
|
||||
Return to the next-step options after sharing.
|
||||
|
||||
#### 6.4 End the Session
|
||||
|
||||
When ending:
|
||||
- offer to commit only the ideation doc
|
||||
- do not create a branch
|
||||
- do not push
|
||||
- if the user declines, leave the file uncommitted
|
||||
|
||||
## Quality Bar
|
||||
|
||||
Before finishing, check:
|
||||
|
||||
- the idea set is grounded in the actual repo
|
||||
- the candidate list was generated before filtering
|
||||
- the original many-ideas -> critique -> survivors mechanism was preserved
|
||||
- if sub-agents were used, they improved diversity without replacing the core workflow
|
||||
- every rejected idea has a reason
|
||||
- survivors are materially better than a naive "give me ideas" list
|
||||
- the artifact was written before any handoff, sharing, or session end
|
||||
- acting on an idea routes to `ce:brainstorm`, not directly to implementation
|
||||
571
plugins/compound-engineering/skills/ce-plan-beta/SKILL.md
Normal file
571
plugins/compound-engineering/skills/ce-plan-beta/SKILL.md
Normal file
@@ -0,0 +1,571 @@
|
||||
---
|
||||
name: ce:plan-beta
|
||||
description: "[BETA] Transform feature descriptions or requirements into structured implementation plans grounded in repo patterns and research. Use when the user says 'plan this', 'create a plan', 'write a tech plan', 'plan the implementation', 'how should we build', 'what's the approach for', 'break this down', or when a brainstorm/requirements document is ready for technical planning. Best when requirements are at least roughly defined; for exploratory or ambiguous requests, prefer ce:brainstorm first."
|
||||
argument-hint: "[feature description, requirements doc path, or improvement idea]"
|
||||
disable-model-invocation: true
|
||||
---
|
||||
|
||||
# Create Technical Plan
|
||||
|
||||
**Note: The current year is 2026.** Use this when dating plans and searching for recent documentation.
|
||||
|
||||
`ce:brainstorm` defines **WHAT** to build. `ce:plan` defines **HOW** to build it. `ce:work` executes the plan.
|
||||
|
||||
This workflow produces a durable implementation plan. It does **not** implement code, run tests, or learn from execution-time results. If the answer depends on changing code and seeing what happens, that belongs in `ce:work`, not here.
|
||||
|
||||
## Interaction Method
|
||||
|
||||
Use the platform's question tool when available. When asking the user a question, prefer the platform's blocking question tool if one exists (`AskUserQuestion` in Claude Code, `request_user_input` in Codex, `ask_user` in Gemini). Otherwise, present numbered options in chat and wait for the user's reply before proceeding.
|
||||
|
||||
Ask one question at a time. Prefer a concise single-select choice when natural options exist.
|
||||
|
||||
## Feature Description
|
||||
|
||||
<feature_description> #$ARGUMENTS </feature_description>
|
||||
|
||||
**If the feature description above is empty, ask the user:** "What would you like to plan? Please describe the feature, bug fix, or improvement you have in mind."
|
||||
|
||||
Do not proceed until you have a clear planning input.
|
||||
|
||||
## Core Principles
|
||||
|
||||
1. **Use requirements as the source of truth** - If `ce:brainstorm` produced a requirements document, planning should build from it rather than re-inventing behavior.
|
||||
2. **Decisions, not code** - Capture approach, boundaries, files, dependencies, risks, and test scenarios. Do not pre-write implementation code or shell command choreography.
|
||||
3. **Research before structuring** - Explore the codebase, institutional learnings, and external guidance when warranted before finalizing the plan.
|
||||
4. **Right-size the artifact** - Small work gets a compact plan. Large work gets more structure. The philosophy stays the same at every depth.
|
||||
5. **Separate planning from execution discovery** - Resolve planning-time questions here. Explicitly defer execution-time unknowns to implementation.
|
||||
6. **Keep the plan portable** - The plan should work as a living document, review artifact, or issue body without embedding tool-specific executor instructions.
|
||||
|
||||
## Plan Quality Bar
|
||||
|
||||
Every plan should contain:
|
||||
- A clear problem frame and scope boundary
|
||||
- Concrete requirements traceability back to the request or origin document
|
||||
- Exact file paths for the work being proposed
|
||||
- Explicit test file paths for feature-bearing implementation units
|
||||
- Decisions with rationale, not just tasks
|
||||
- Existing patterns or code references to follow
|
||||
- Specific test scenarios and verification outcomes
|
||||
- Clear dependencies and sequencing
|
||||
|
||||
A plan is ready when an implementer can start confidently without needing the plan to write the code for them.
|
||||
|
||||
## Workflow
|
||||
|
||||
### Phase 0: Resume, Source, and Scope
|
||||
|
||||
#### 0.1 Resume Existing Plan Work When Appropriate
|
||||
|
||||
If the user references an existing plan file or there is an obvious recent matching plan in `docs/plans/`:
|
||||
- Read it
|
||||
- Confirm whether to update it in place or create a new plan
|
||||
- If updating, preserve completed checkboxes and revise only the still-relevant sections
|
||||
|
||||
#### 0.2 Find Upstream Requirements Document
|
||||
|
||||
Before asking planning questions, search `docs/brainstorms/` for files matching `*-requirements.md`.
|
||||
|
||||
**Relevance criteria:** A requirements document is relevant if:
|
||||
- The topic semantically matches the feature description
|
||||
- It was created within the last 30 days (use judgment to override if the document is clearly still relevant or clearly stale)
|
||||
- It appears to cover the same user problem or scope
|
||||
|
||||
If multiple source documents match, ask which one to use using the platform's blocking question tool when available (see Interaction Method). Otherwise, present numbered options in chat and wait for the user's reply before proceeding.
|
||||
|
||||
#### 0.3 Use the Source Document as Primary Input
|
||||
|
||||
If a relevant requirements document exists:
|
||||
1. Read it thoroughly
|
||||
2. Announce that it will serve as the origin document for planning
|
||||
3. Carry forward all of the following:
|
||||
- Problem frame
|
||||
- Requirements and success criteria
|
||||
- Scope boundaries
|
||||
- Key decisions and rationale
|
||||
- Dependencies or assumptions
|
||||
- Outstanding questions, preserving whether they are blocking or deferred
|
||||
4. Use the source document as the primary input to planning and research
|
||||
5. Reference important carried-forward decisions in the plan with `(see origin: <source-path>)`
|
||||
6. Do not silently omit source content — if the origin document discussed it, the plan must address it even if briefly. Before finalizing, scan each section of the origin document to verify nothing was dropped.
|
||||
|
||||
If no relevant requirements document exists, planning may proceed from the user's request directly.
|
||||
|
||||
#### 0.4 No-Requirements-Doc Fallback
|
||||
|
||||
If no relevant requirements document exists:
|
||||
- Assess whether the request is already clear enough for direct technical planning
|
||||
- If the ambiguity is mainly product framing, user behavior, or scope definition, recommend `ce:brainstorm` first
|
||||
- If the user wants to continue here anyway, run a short planning bootstrap instead of refusing
|
||||
|
||||
The planning bootstrap should establish:
|
||||
- Problem frame
|
||||
- Intended behavior
|
||||
- Scope boundaries and obvious non-goals
|
||||
- Success criteria
|
||||
- Blocking questions or assumptions
|
||||
|
||||
Keep this bootstrap brief. It exists to preserve direct-entry convenience, not to replace a full brainstorm.
|
||||
|
||||
If the bootstrap uncovers major unresolved product questions:
|
||||
- Recommend `ce:brainstorm` again
|
||||
- If the user still wants to continue, require explicit assumptions before proceeding
|
||||
|
||||
#### 0.5 Classify Outstanding Questions Before Planning
|
||||
|
||||
If the origin document contains `Resolve Before Planning` or similar blocking questions:
|
||||
- Review each one before proceeding
|
||||
- Reclassify it into planning-owned work **only if** it is actually a technical, architectural, or research question
|
||||
- Keep it as a blocker if it would change product behavior, scope, or success criteria
|
||||
|
||||
If true product blockers remain:
|
||||
- Surface them clearly
|
||||
- Ask the user, using the platform's blocking question tool when available (see Interaction Method), whether to:
|
||||
1. Resume `ce:brainstorm` to resolve them
|
||||
2. Convert them into explicit assumptions or decisions and continue
|
||||
- Do not continue planning while true blockers remain unresolved
|
||||
|
||||
#### 0.6 Assess Plan Depth
|
||||
|
||||
Classify the work into one of these plan depths:
|
||||
|
||||
- **Lightweight** - small, well-bounded, low ambiguity
|
||||
- **Standard** - normal feature or bounded refactor with some technical decisions to document
|
||||
- **Deep** - cross-cutting, strategic, high-risk, or highly ambiguous implementation work
|
||||
|
||||
If depth is unclear, ask one targeted question and then continue.
|
||||
|
||||
### Phase 1: Gather Context
|
||||
|
||||
#### 1.1 Local Research (Always Runs)
|
||||
|
||||
Prepare a concise planning context summary (a paragraph or two) to pass as input to the research agents:
|
||||
- If an origin document exists, summarize the problem frame, requirements, and key decisions from that document
|
||||
- Otherwise use the feature description directly
|
||||
|
||||
Run these agents in parallel:
|
||||
|
||||
- Task compound-engineering:research:repo-research-analyst(planning context summary)
|
||||
- Task compound-engineering:research:learnings-researcher(planning context summary)
|
||||
|
||||
Collect:
|
||||
- Existing patterns and conventions to follow
|
||||
- Relevant files, modules, and tests
|
||||
- AGENTS.md guidance that materially affects the plan, with CLAUDE.md used only as compatibility fallback when present
|
||||
- Institutional learnings from `docs/solutions/`
|
||||
|
||||
#### 1.2 Decide on External Research
|
||||
|
||||
Based on the origin document, user signals, and local findings, decide whether external research adds value.
|
||||
|
||||
**Read between the lines.** Pay attention to signals from the conversation so far:
|
||||
- **User familiarity** — Are they pointing to specific files or patterns? They likely know the codebase well.
|
||||
- **User intent** — Do they want speed or thoroughness? Exploration or execution?
|
||||
- **Topic risk** — Security, payments, external APIs warrant more caution regardless of user signals.
|
||||
- **Uncertainty level** — Is the approach clear or still open-ended?
|
||||
|
||||
**Always lean toward external research when:**
|
||||
- The topic is high-risk: security, payments, privacy, external APIs, migrations, compliance
|
||||
- The codebase lacks relevant local patterns
|
||||
- The user is exploring unfamiliar territory
|
||||
|
||||
**Skip external research when:**
|
||||
- The codebase already shows a strong local pattern
|
||||
- The user already knows the intended shape
|
||||
- Additional external context would add little practical value
|
||||
|
||||
Announce the decision briefly before continuing. Examples:
|
||||
- "Your codebase has solid patterns for this. Proceeding without external research."
|
||||
- "This involves payment processing, so I'll research current best practices first."
|
||||
|
||||
#### 1.3 External Research (Conditional)
|
||||
|
||||
If Step 1.2 indicates external research is useful, run these agents in parallel:
|
||||
|
||||
- Task compound-engineering:research:best-practices-researcher(planning context summary)
|
||||
- Task compound-engineering:research:framework-docs-researcher(planning context summary)
|
||||
|
||||
#### 1.4 Consolidate Research
|
||||
|
||||
Summarize:
|
||||
- Relevant codebase patterns and file paths
|
||||
- Relevant institutional learnings
|
||||
- External references and best practices, if gathered
|
||||
- Related issues, PRs, or prior art
|
||||
- Any constraints that should materially shape the plan
|
||||
|
||||
#### 1.5 Flow and Edge-Case Analysis (Conditional)
|
||||
|
||||
For **Standard** or **Deep** plans, or when user flow completeness is still unclear, run:
|
||||
|
||||
- Task compound-engineering:workflow:spec-flow-analyzer(planning context summary, research findings)
|
||||
|
||||
Use the output to:
|
||||
- Identify missing edge cases, state transitions, or handoff gaps
|
||||
- Tighten requirements trace or verification strategy
|
||||
- Add only the flow details that materially improve the plan
|
||||
|
||||
### Phase 2: Resolve Planning Questions
|
||||
|
||||
Build a planning question list from:
|
||||
- Deferred questions in the origin document
|
||||
- Gaps discovered in repo or external research
|
||||
- Technical decisions required to produce a useful plan
|
||||
|
||||
For each question, decide whether it should be:
|
||||
- **Resolved during planning** - the answer is knowable from repo context, documentation, or user choice
|
||||
- **Deferred to implementation** - the answer depends on code changes, runtime behavior, or execution-time discovery
|
||||
|
||||
Ask the user only when the answer materially affects architecture, scope, sequencing, or risk and cannot be responsibly inferred. Use the platform's blocking question tool when available (see Interaction Method).
|
||||
|
||||
**Do not** run tests, build the app, or probe runtime behavior in this phase. The goal is a strong plan, not partial execution.
|
||||
|
||||
### Phase 3: Structure the Plan
|
||||
|
||||
#### 3.1 Title and File Naming
|
||||
|
||||
- Draft a clear, searchable title using conventional format such as `feat: Add user authentication` or `fix: Prevent checkout double-submit`
|
||||
- Determine the plan type: `feat`, `fix`, or `refactor`
|
||||
- Build the filename following the repository convention: `docs/plans/YYYY-MM-DD-NNN-<type>-<descriptive-name>-beta-plan.md`
|
||||
- Create `docs/plans/` if it does not exist
|
||||
- Check existing files for today's date to determine the next sequence number (zero-padded to 3 digits, starting at 001)
|
||||
- Keep the descriptive name concise (3-5 words) and kebab-cased
|
||||
- Append `-beta` before `-plan` to distinguish from stable-generated plans
|
||||
- Examples: `2026-01-15-001-feat-user-authentication-flow-beta-plan.md`, `2026-02-03-002-fix-checkout-race-condition-beta-plan.md`
|
||||
- Avoid: missing sequence numbers, vague names like "new-feature", invalid characters (colons, spaces)
|
||||
|
||||
#### 3.2 Stakeholder and Impact Awareness
|
||||
|
||||
For **Standard** or **Deep** plans, briefly consider who is affected by this change — end users, developers, operations, other teams — and how that should shape the plan. For cross-cutting work, note affected parties in the System-Wide Impact section.
|
||||
|
||||
#### 3.3 Break Work into Implementation Units
|
||||
|
||||
Break the work into logical implementation units. Each unit should represent one meaningful change that an implementer could typically land as an atomic commit.
|
||||
|
||||
Good units are:
|
||||
- Focused on one component, behavior, or integration seam
|
||||
- Usually touching a small cluster of related files
|
||||
- Ordered by dependency
|
||||
- Concrete enough for execution without pre-writing code
|
||||
- Marked with checkbox syntax for progress tracking
|
||||
|
||||
Avoid:
|
||||
- 2-5 minute micro-steps
|
||||
- Units that span multiple unrelated concerns
|
||||
- Units that are so vague an implementer still has to invent the plan
|
||||
|
||||
#### 3.4 Define Each Implementation Unit
|
||||
|
||||
For each unit, include:
|
||||
- **Goal** - what this unit accomplishes
|
||||
- **Requirements** - which requirements or success criteria it advances
|
||||
- **Dependencies** - what must exist first
|
||||
- **Files** - exact file paths to create, modify, or test
|
||||
- **Approach** - key decisions, data flow, component boundaries, or integration notes
|
||||
- **Patterns to follow** - existing code or conventions to mirror
|
||||
- **Test scenarios** - specific behaviors, edge cases, and failure paths to cover
|
||||
- **Verification** - how an implementer should know the unit is complete, expressed as outcomes rather than shell command scripts
|
||||
|
||||
Every feature-bearing unit should include the test file path in `**Files:**`.
|
||||
|
||||
#### 3.5 Keep Planning-Time and Implementation-Time Unknowns Separate
|
||||
|
||||
If something is important but not knowable yet, record it explicitly under deferred implementation notes rather than pretending to resolve it in the plan.
|
||||
|
||||
Examples:
|
||||
- Exact method or helper names
|
||||
- Final SQL or query details after touching real code
|
||||
- Runtime behavior that depends on seeing actual test failures
|
||||
- Refactors that may become unnecessary once implementation starts
|
||||
|
||||
### Phase 4: Write the Plan
|
||||
|
||||
Use one planning philosophy across all depths. Change the amount of detail, not the boundary between planning and execution.
|
||||
|
||||
#### 4.1 Plan Depth Guidance
|
||||
|
||||
**Lightweight**
|
||||
- Keep the plan compact
|
||||
- Usually 2-4 implementation units
|
||||
- Omit optional sections that add little value
|
||||
|
||||
**Standard**
|
||||
- Use the full core template
|
||||
- Usually 3-6 implementation units
|
||||
- Include risks, deferred questions, and system-wide impact when relevant
|
||||
|
||||
**Deep**
|
||||
- Use the full core template plus optional analysis sections
|
||||
- Usually 4-8 implementation units
|
||||
- Group units into phases when that improves clarity
|
||||
- Include alternatives considered, documentation impacts, and deeper risk treatment when warranted
|
||||
|
||||
#### 4.1b Optional Deep Plan Extensions
|
||||
|
||||
For sufficiently large, risky, or cross-cutting work, add the sections that genuinely help:
|
||||
- **Alternative Approaches Considered**
|
||||
- **Success Metrics**
|
||||
- **Dependencies / Prerequisites**
|
||||
- **Risk Analysis & Mitigation**
|
||||
- **Phased Delivery**
|
||||
- **Documentation Plan**
|
||||
- **Operational / Rollout Notes**
|
||||
- **Future Considerations** only when they materially affect current design
|
||||
|
||||
Do not add these as boilerplate. Include them only when they improve execution quality or stakeholder alignment.
|
||||
|
||||
#### 4.2 Core Plan Template
|
||||
|
||||
Omit clearly inapplicable optional sections, especially for Lightweight plans.
|
||||
|
||||
```markdown
|
||||
---
|
||||
title: [Plan Title]
|
||||
type: [feat|fix|refactor]
|
||||
status: active
|
||||
date: YYYY-MM-DD
|
||||
origin: docs/brainstorms/YYYY-MM-DD-<topic>-requirements.md # include when planning from a requirements doc
|
||||
deepened: YYYY-MM-DD # optional, set later by deepen-plan-beta when the plan is substantively strengthened
|
||||
---
|
||||
|
||||
# [Plan Title]
|
||||
|
||||
## Overview
|
||||
|
||||
[What is changing and why]
|
||||
|
||||
## Problem Frame
|
||||
|
||||
[Summarize the user/business problem and context. Reference the origin doc when present.]
|
||||
|
||||
## Requirements Trace
|
||||
|
||||
- R1. [Requirement or success criterion this plan must satisfy]
|
||||
- R2. [Requirement or success criterion this plan must satisfy]
|
||||
|
||||
## Scope Boundaries
|
||||
|
||||
- [Explicit non-goal or exclusion]
|
||||
|
||||
## Context & Research
|
||||
|
||||
### Relevant Code and Patterns
|
||||
|
||||
- [Existing file, class, component, or pattern to follow]
|
||||
|
||||
### Institutional Learnings
|
||||
|
||||
- [Relevant `docs/solutions/` insight]
|
||||
|
||||
### External References
|
||||
|
||||
- [Relevant external docs or best-practice source, if used]
|
||||
|
||||
## Key Technical Decisions
|
||||
|
||||
- [Decision]: [Rationale]
|
||||
|
||||
## Open Questions
|
||||
|
||||
### Resolved During Planning
|
||||
|
||||
- [Question]: [Resolution]
|
||||
|
||||
### Deferred to Implementation
|
||||
|
||||
- [Question or unknown]: [Why it is intentionally deferred]
|
||||
|
||||
## Implementation Units
|
||||
|
||||
- [ ] **Unit 1: [Name]**
|
||||
|
||||
**Goal:** [What this unit accomplishes]
|
||||
|
||||
**Requirements:** [R1, R2]
|
||||
|
||||
**Dependencies:** [None / Unit 1 / external prerequisite]
|
||||
|
||||
**Files:**
|
||||
- Create: `path/to/new_file`
|
||||
- Modify: `path/to/existing_file`
|
||||
- Test: `path/to/test_file`
|
||||
|
||||
**Approach:**
|
||||
- [Key design or sequencing decision]
|
||||
|
||||
**Patterns to follow:**
|
||||
- [Existing file, class, or pattern]
|
||||
|
||||
**Test scenarios:**
|
||||
- [Specific scenario with expected behavior]
|
||||
- [Edge case or failure path]
|
||||
|
||||
**Verification:**
|
||||
- [Outcome that should hold when this unit is complete]
|
||||
|
||||
## System-Wide Impact
|
||||
|
||||
- **Interaction graph:** [What callbacks, middleware, observers, or entry points may be affected]
|
||||
- **Error propagation:** [How failures should travel across layers]
|
||||
- **State lifecycle risks:** [Partial-write, cache, duplicate, or cleanup concerns]
|
||||
- **API surface parity:** [Other interfaces that may require the same change]
|
||||
- **Integration coverage:** [Cross-layer scenarios unit tests alone will not prove]
|
||||
|
||||
## Risks & Dependencies
|
||||
|
||||
- [Meaningful risk, dependency, or sequencing concern]
|
||||
|
||||
## Documentation / Operational Notes
|
||||
|
||||
- [Docs, rollout, monitoring, or support impacts when relevant]
|
||||
|
||||
## Sources & References
|
||||
|
||||
- **Origin document:** [docs/brainstorms/YYYY-MM-DD-<topic>-requirements.md](path)
|
||||
- Related code: [path or symbol]
|
||||
- Related PRs/issues: #[number]
|
||||
- External docs: [url]
|
||||
```
|
||||
|
||||
For larger `Deep` plans, extend the core template only when useful with sections such as:
|
||||
|
||||
```markdown
|
||||
## Alternative Approaches Considered
|
||||
|
||||
- [Approach]: [Why rejected or not chosen]
|
||||
|
||||
## Success Metrics
|
||||
|
||||
- [How we will know this solved the intended problem]
|
||||
|
||||
## Dependencies / Prerequisites
|
||||
|
||||
- [Technical, organizational, or rollout dependency]
|
||||
|
||||
## Risk Analysis & Mitigation
|
||||
|
||||
- [Risk]: [Mitigation]
|
||||
|
||||
## Phased Delivery
|
||||
|
||||
### Phase 1
|
||||
- [What lands first and why]
|
||||
|
||||
### Phase 2
|
||||
- [What follows and why]
|
||||
|
||||
## Documentation Plan
|
||||
|
||||
- [Docs or runbooks to update]
|
||||
|
||||
## Operational / Rollout Notes
|
||||
|
||||
- [Monitoring, migration, feature flag, or rollout considerations]
|
||||
```
|
||||
|
||||
#### 4.3 Planning Rules
|
||||
|
||||
- Prefer path plus class/component/pattern references over brittle line numbers
|
||||
- Keep implementation units checkable with `- [ ]` syntax for progress tracking
|
||||
- Do not include fenced implementation code blocks unless the plan itself is about code shape as a design artifact
|
||||
- Do not include git commands, commit messages, or exact test command recipes
|
||||
- Do not pretend an execution-time question is settled just to make the plan look complete
|
||||
- Include mermaid diagrams when they clarify relationships or flows that prose alone would make hard to follow — ERDs for data model changes, sequence diagrams for multi-service interactions, state diagrams for lifecycle transitions, flowcharts for complex branching logic
|
||||
|
||||
### Phase 5: Final Review, Write File, and Handoff
|
||||
|
||||
#### 5.1 Review Before Writing
|
||||
|
||||
Before finalizing, check:
|
||||
- The plan does not invent product behavior that should have been defined in `ce:brainstorm`
|
||||
- If there was no origin document, the bounded planning bootstrap established enough product clarity to plan responsibly
|
||||
- Every major decision is grounded in the origin document or research
|
||||
- Each implementation unit is concrete, dependency-ordered, and implementation-ready
|
||||
- Test scenarios are specific without becoming test code
|
||||
- Deferred items are explicit and not hidden as fake certainty
|
||||
|
||||
If the plan originated from a requirements document, re-read that document and verify:
|
||||
- The chosen approach still matches the product intent
|
||||
- Scope boundaries and success criteria are preserved
|
||||
- Blocking questions were either resolved, explicitly assumed, or sent back to `ce:brainstorm`
|
||||
- Every section of the origin document is addressed in the plan — scan each section to confirm nothing was silently dropped
|
||||
|
||||
#### 5.2 Write Plan File
|
||||
|
||||
**REQUIRED: Write the plan file to disk before presenting any options.**
|
||||
|
||||
Use the Write tool to save the complete plan to:
|
||||
|
||||
```text
|
||||
docs/plans/YYYY-MM-DD-NNN-<type>-<descriptive-name>-beta-plan.md
|
||||
```
|
||||
|
||||
Confirm:
|
||||
|
||||
```text
|
||||
Plan written to docs/plans/[filename]
|
||||
```
|
||||
|
||||
**Pipeline mode:** If invoked from an automated workflow such as LFG, SLFG, or any `disable-model-invocation` context, skip interactive questions. Make the needed choices automatically and proceed to writing the plan.
|
||||
|
||||
#### 5.3 Post-Generation Options
|
||||
|
||||
After writing the plan file, present the options using the platform's blocking question tool when available (see Interaction Method). Otherwise present numbered options in chat and wait for the user's reply before proceeding.
|
||||
|
||||
**Question:** "Plan ready at `docs/plans/YYYY-MM-DD-NNN-<type>-<name>-beta-plan.md`. What would you like to do next?"
|
||||
|
||||
**Options:**
|
||||
1. **Open plan in editor** - Open the plan file for review
|
||||
2. **Run `/deepen-plan-beta`** - Stress-test weak sections with targeted research when the plan needs more confidence
|
||||
3. **Run `document-review` skill** - Improve the plan through structured document review
|
||||
4. **Share to Proof** - Upload the plan for collaborative review and sharing
|
||||
5. **Start `/ce:work`** - Begin implementing this plan in the current environment
|
||||
6. **Start `/ce:work` in another session** - Begin implementing in a separate agent session when the current platform supports it
|
||||
7. **Create Issue** - Create an issue in the configured tracker
|
||||
|
||||
Based on selection:
|
||||
- **Open plan in editor** → Open `docs/plans/<plan_filename>.md` using the current platform's file-open or editor mechanism (e.g., `open` on macOS, `xdg-open` on Linux, or the IDE's file-open API)
|
||||
- **`/deepen-plan-beta`** → Call `/deepen-plan-beta` with the plan path
|
||||
- **`document-review` skill** → Load the `document-review` skill with the plan path
|
||||
- **Share to Proof** → Upload the plan:
|
||||
```bash
|
||||
CONTENT=$(cat docs/plans/<plan_filename>.md)
|
||||
TITLE="Plan: <plan title from frontmatter>"
|
||||
RESPONSE=$(curl -s -X POST https://www.proofeditor.ai/share/markdown \
|
||||
-H "Content-Type: application/json" \
|
||||
-d "$(jq -n --arg title "$TITLE" --arg markdown "$CONTENT" --arg by "ai:compound" '{title: $title, markdown: $markdown, by: $by}')")
|
||||
PROOF_URL=$(echo "$RESPONSE" | jq -r '.tokenUrl')
|
||||
```
|
||||
Display `View & collaborate in Proof: <PROOF_URL>` if successful, then return to the options
|
||||
- **`/ce:work`** → Call `/ce:work` with the plan path
|
||||
- **`/ce:work` in another session** → If the current platform supports launching a separate agent session, start `/ce:work` with the plan path there. Otherwise, explain the limitation briefly and offer to run `/ce:work` in the current session instead.
|
||||
- **Create Issue** → Follow the Issue Creation section below
|
||||
- **Other** → Accept free text for revisions and loop back to options
|
||||
|
||||
If running with ultrathink enabled, or the platform's reasoning/effort level is set to max or extra-high, automatically run `/deepen-plan-beta` only when the plan is `Standard` or `Deep`, high-risk, or still shows meaningful confidence gaps in decisions, sequencing, system-wide impact, risks, or verification.
|
||||
|
||||
## Issue Creation
|
||||
|
||||
When the user selects "Create Issue", detect their project tracker from `AGENTS.md` or, if needed for compatibility, `CLAUDE.md`:
|
||||
|
||||
1. Look for `project_tracker: github` or `project_tracker: linear`
|
||||
2. If GitHub:
|
||||
|
||||
```bash
|
||||
gh issue create --title "<type>: <title>" --body-file <plan_path>
|
||||
```
|
||||
|
||||
3. If Linear:
|
||||
|
||||
```bash
|
||||
linear issue create --title "<title>" --description "$(cat <plan_path>)"
|
||||
```
|
||||
|
||||
4. If no tracker is configured:
|
||||
- Ask which tracker they use using the platform's blocking question tool when available (see Interaction Method)
|
||||
- Suggest adding the tracker to `AGENTS.md` for future runs
|
||||
|
||||
After issue creation:
|
||||
- Display the issue URL
|
||||
- Ask whether to proceed to `/ce:work`
|
||||
|
||||
NEVER CODE! Research, decide, and write the plan.
|
||||
@@ -1,5 +1,5 @@
|
||||
---
|
||||
name: workflows:plan
|
||||
name: ce:plan
|
||||
description: Transform feature descriptions into well-structured project plans following conventions
|
||||
argument-hint: "[feature description, bug report, or improvement idea]"
|
||||
---
|
||||
@@ -22,30 +22,39 @@ Do not proceed until you have a clear feature description from the user.
|
||||
|
||||
### 0. Idea Refinement
|
||||
|
||||
**Check for brainstorm output first:**
|
||||
**Check for requirements document first:**
|
||||
|
||||
Before asking questions, look for recent brainstorm documents in `docs/brainstorms/` that match this feature:
|
||||
Before asking questions, look for recent requirements documents in `docs/brainstorms/` that match this feature:
|
||||
|
||||
```bash
|
||||
ls -la docs/brainstorms/*.md 2>/dev/null | head -10
|
||||
ls -la docs/brainstorms/*-requirements.md 2>/dev/null | head -10
|
||||
```
|
||||
|
||||
**Relevance criteria:** A brainstorm is relevant if:
|
||||
**Relevance criteria:** A requirements document is relevant if:
|
||||
- The topic (from filename or YAML frontmatter) semantically matches the feature description
|
||||
- Created within the last 14 days
|
||||
- If multiple candidates match, use the most recent one
|
||||
|
||||
**If a relevant brainstorm exists:**
|
||||
1. Read the brainstorm document
|
||||
2. Announce: "Found brainstorm from [date]: [topic]. Using as context for planning."
|
||||
3. Extract key decisions, chosen approach, and open questions
|
||||
4. **Skip the idea refinement questions below** - the brainstorm already answered WHAT to build
|
||||
5. Use brainstorm decisions as input to the research phase
|
||||
**If a relevant requirements document exists:**
|
||||
1. Read the source document **thoroughly** — every section matters
|
||||
2. Announce: "Found source document from [date]: [topic]. Using as foundation for planning."
|
||||
3. Extract and carry forward **ALL** of the following into the plan:
|
||||
- Key decisions and their rationale
|
||||
- Chosen approach and why alternatives were rejected
|
||||
- Problem framing, constraints, and requirements captured during brainstorming
|
||||
- Outstanding questions, preserving whether they block planning or are intentionally deferred
|
||||
- Success criteria and scope boundaries
|
||||
- Dependencies and assumptions, plus any high-level technical direction only when the origin document is inherently technical
|
||||
4. **Skip the idea refinement questions below** — the source document already answered WHAT to build
|
||||
5. Use source document content as the **primary input** to research and planning phases
|
||||
6. **Critical: The source document is the origin document.** Throughout the plan, reference specific decisions with `(see origin: <source-path>)` when carrying forward conclusions. Do not paraphrase decisions in a way that loses their original context — link back to the source.
|
||||
7. **Do not omit source content** — if the source document discussed it, the plan must address it (even if briefly). Scan each section before finalizing the plan to verify nothing was dropped.
|
||||
8. **If `Resolve Before Planning` contains any items, stop.** Do not proceed with planning. Tell the user planning is blocked by unanswered brainstorm questions and direct them to resume `/ce:brainstorm` or answer those questions first.
|
||||
|
||||
**If multiple brainstorms could match:**
|
||||
Use **AskUserQuestion tool** to ask which brainstorm to use, or whether to proceed without one.
|
||||
**If multiple source documents could match:**
|
||||
Use **AskUserQuestion tool** to ask which source document to use, or whether to proceed without one.
|
||||
|
||||
**If no brainstorm found (or not relevant), run idea refinement:**
|
||||
**If no requirements document is found (or not relevant), run idea refinement:**
|
||||
|
||||
Refine the idea through collaborative dialogue using the **AskUserQuestion tool**:
|
||||
|
||||
@@ -74,11 +83,11 @@ First, I need to understand the project's conventions, existing patterns, and an
|
||||
|
||||
Run these agents **in parallel** to gather local context:
|
||||
|
||||
- Task repo-research-analyst(feature_description)
|
||||
- Task learnings-researcher(feature_description)
|
||||
- Task compound-engineering:research:repo-research-analyst(feature_description)
|
||||
- Task compound-engineering:research:learnings-researcher(feature_description)
|
||||
|
||||
**What to look for:**
|
||||
- **Repo research:** existing patterns, CLAUDE.md guidance, technology familiarity, pattern consistency
|
||||
- **Repo research:** existing patterns, AGENTS.md guidance, technology familiarity, pattern consistency
|
||||
- **Learnings:** documented solutions in `docs/solutions/` that might apply (gotchas, patterns, lessons learned)
|
||||
|
||||
These findings inform the next step.
|
||||
@@ -89,7 +98,7 @@ Based on signals from Step 0 and findings from Step 1, decide on external resear
|
||||
|
||||
**High-risk topics → always research.** Security, payments, external APIs, data privacy. The cost of missing something is too high. This takes precedence over speed signals.
|
||||
|
||||
**Strong local context → skip external research.** Codebase has good patterns, CLAUDE.md has guidance, user knows what they want. External research adds little value.
|
||||
**Strong local context -> skip external research.** Codebase has good patterns, AGENTS.md has guidance, user knows what they want. External research adds little value.
|
||||
|
||||
**Uncertainty or unfamiliar territory → research.** User is exploring, codebase has no examples, new technology. External perspective is valuable.
|
||||
|
||||
@@ -105,8 +114,8 @@ Examples:
|
||||
|
||||
Run these agents in parallel:
|
||||
|
||||
- Task best-practices-researcher(feature_description)
|
||||
- Task framework-docs-researcher(feature_description)
|
||||
- Task compound-engineering:research:best-practices-researcher(feature_description)
|
||||
- Task compound-engineering:research:framework-docs-researcher(feature_description)
|
||||
|
||||
### 1.6. Consolidate Research
|
||||
|
||||
@@ -116,7 +125,7 @@ After all research steps complete, consolidate findings:
|
||||
- **Include relevant institutional learnings** from `docs/solutions/` (key insights, gotchas to avoid)
|
||||
- Note external documentation URLs and best practices (if external research was done)
|
||||
- List related issues or PRs discovered
|
||||
- Capture CLAUDE.md conventions
|
||||
- Capture AGENTS.md conventions
|
||||
|
||||
**Optional validation:** Briefly summarize findings and ask if anything looks off or missing before proceeding to planning.
|
||||
|
||||
@@ -130,8 +139,11 @@ Think like a product manager - what would make this issue clear and actionable?
|
||||
|
||||
- [ ] Draft clear, searchable issue title using conventional format (e.g., `feat: Add user authentication`, `fix: Cart total calculation`)
|
||||
- [ ] Determine issue type: enhancement, bug, refactor
|
||||
- [ ] Convert title to filename: add today's date prefix, strip prefix colon, kebab-case, add `-plan` suffix
|
||||
- Example: `feat: Add User Authentication` → `2026-01-21-feat-add-user-authentication-plan.md`
|
||||
- [ ] Convert title to filename: add today's date prefix, determine daily sequence number, strip prefix colon, kebab-case, add `-plan` suffix
|
||||
- Scan `docs/plans/` for files matching today's date pattern `YYYY-MM-DD-\d{3}-`
|
||||
- Find the highest existing sequence number for today
|
||||
- Increment by 1, zero-padded to 3 digits (001, 002, etc.)
|
||||
- Example: `feat: Add User Authentication` → `2026-01-21-001-feat-add-user-authentication-plan.md`
|
||||
- Keep it descriptive (3-5 words after prefix) so plans are findable by context
|
||||
|
||||
**Stakeholder Analysis:**
|
||||
@@ -150,7 +162,7 @@ Think like a product manager - what would make this issue clear and actionable?
|
||||
|
||||
After planning the issue structure, run SpecFlow Analyzer to validate and refine the feature specification:
|
||||
|
||||
- Task spec-flow-analyzer(feature_description, research_findings)
|
||||
- Task compound-engineering:workflow:spec-flow-analyzer(feature_description, research_findings)
|
||||
|
||||
**SpecFlow Analyzer Output:**
|
||||
|
||||
@@ -180,6 +192,7 @@ title: [Issue Title]
|
||||
type: [feat|fix|refactor]
|
||||
status: active
|
||||
date: YYYY-MM-DD
|
||||
origin: docs/brainstorms/YYYY-MM-DD-<topic>-requirements.md # if originated from a requirements doc, otherwise omit
|
||||
---
|
||||
|
||||
# [Issue Title]
|
||||
@@ -207,8 +220,9 @@ class Test
|
||||
end
|
||||
```
|
||||
|
||||
## References
|
||||
## Sources
|
||||
|
||||
- **Origin document:** [docs/brainstorms/YYYY-MM-DD-<topic>-requirements.md](path) — include if plan originated from an upstream requirements doc
|
||||
- Related issue: #[issue_number]
|
||||
- Documentation: [relevant_docs_url]
|
||||
````
|
||||
@@ -233,6 +247,7 @@ title: [Issue Title]
|
||||
type: [feat|fix|refactor]
|
||||
status: active
|
||||
date: YYYY-MM-DD
|
||||
origin: docs/brainstorms/YYYY-MM-DD-<topic>-requirements.md # if originated from a requirements doc, otherwise omit
|
||||
---
|
||||
|
||||
# [Issue Title]
|
||||
@@ -255,6 +270,14 @@ date: YYYY-MM-DD
|
||||
- Performance implications
|
||||
- Security considerations
|
||||
|
||||
## System-Wide Impact
|
||||
|
||||
- **Interaction graph**: [What callbacks/middleware/observers fire when this runs?]
|
||||
- **Error propagation**: [How do errors flow across layers? Do retry strategies align?]
|
||||
- **State lifecycle risks**: [Can partial failure leave orphaned/inconsistent state?]
|
||||
- **API surface parity**: [What other interfaces expose similar functionality and need the same change?]
|
||||
- **Integration test scenarios**: [Cross-layer scenarios that unit tests won't catch]
|
||||
|
||||
## Acceptance Criteria
|
||||
|
||||
- [ ] Detailed requirement 1
|
||||
@@ -269,8 +292,9 @@ date: YYYY-MM-DD
|
||||
|
||||
[What could block or complicate this]
|
||||
|
||||
## References & Research
|
||||
## Sources & References
|
||||
|
||||
- **Origin document:** [docs/brainstorms/YYYY-MM-DD-<topic>-requirements.md](path) — include if plan originated from an upstream requirements doc
|
||||
- Similar implementations: [file_path:line_number]
|
||||
- Best practices: [documentation_url]
|
||||
- Related PRs: #[pr_number]
|
||||
@@ -298,6 +322,7 @@ title: [Issue Title]
|
||||
type: [feat|fix|refactor]
|
||||
status: active
|
||||
date: YYYY-MM-DD
|
||||
origin: docs/brainstorms/YYYY-MM-DD-<topic>-requirements.md # if originated from a requirements doc, otherwise omit
|
||||
---
|
||||
|
||||
# [Issue Title]
|
||||
@@ -344,6 +369,28 @@ date: YYYY-MM-DD
|
||||
|
||||
[Other solutions evaluated and why rejected]
|
||||
|
||||
## System-Wide Impact
|
||||
|
||||
### Interaction Graph
|
||||
|
||||
[Map the chain reaction: what callbacks, middleware, observers, and event handlers fire when this code runs? Trace at least two levels deep. Document: "Action X triggers Y, which calls Z, which persists W."]
|
||||
|
||||
### Error & Failure Propagation
|
||||
|
||||
[Trace errors from lowest layer up. List specific error classes and where they're handled. Identify retry conflicts, unhandled error types, and silent failure swallowing.]
|
||||
|
||||
### State Lifecycle Risks
|
||||
|
||||
[Walk through each step that persists state. Can partial failure orphan rows, duplicate records, or leave caches stale? Document cleanup mechanisms or their absence.]
|
||||
|
||||
### API Surface Parity
|
||||
|
||||
[List all interfaces (classes, DSLs, endpoints) that expose equivalent functionality. Note which need updating and which share the code path.]
|
||||
|
||||
### Integration Test Scenarios
|
||||
|
||||
[3-5 cross-layer test scenarios that unit tests with mocks would never catch. Include expected behavior for each.]
|
||||
|
||||
## Acceptance Criteria
|
||||
|
||||
### Functional Requirements
|
||||
@@ -386,7 +433,11 @@ date: YYYY-MM-DD
|
||||
|
||||
[What docs need updating]
|
||||
|
||||
## References & Research
|
||||
## Sources & References
|
||||
|
||||
### Origin
|
||||
|
||||
- **Origin document:** [docs/brainstorms/YYYY-MM-DD-<topic>-requirements.md](path) — include if plan originated from an upstream requirements doc. Key decisions carried forward: [list 2-3 major decisions from the origin]
|
||||
|
||||
### Internal References
|
||||
|
||||
@@ -465,6 +516,16 @@ end
|
||||
|
||||
### 6. Final Review & Submission
|
||||
|
||||
**Origin document cross-check (if plan originated from a requirements doc):**
|
||||
|
||||
Before finalizing, re-read the origin document and verify:
|
||||
- [ ] Every key decision from the origin document is reflected in the plan
|
||||
- [ ] The chosen approach matches what was decided in the origin document
|
||||
- [ ] Constraints and requirements from the origin document are captured in acceptance criteria
|
||||
- [ ] Open questions from the origin document are either resolved or flagged
|
||||
- [ ] The `origin:` frontmatter field points to the correct source file
|
||||
- [ ] The Sources section includes the origin document with a summary of carried-forward decisions
|
||||
|
||||
**Pre-submission Checklist:**
|
||||
|
||||
- [ ] Title is searchable and descriptive
|
||||
@@ -475,57 +536,84 @@ end
|
||||
- [ ] Add names of files in pseudo code examples and todo lists
|
||||
- [ ] Add an ERD mermaid diagram if applicable for new model changes
|
||||
|
||||
## Write Plan File
|
||||
|
||||
**REQUIRED: Write the plan file to disk before presenting any options.**
|
||||
|
||||
```bash
|
||||
mkdir -p docs/plans/
|
||||
# Determine daily sequence number
|
||||
today=$(date +%Y-%m-%d)
|
||||
last_seq=$(ls docs/plans/${today}-*-plan.md 2>/dev/null | grep -oP "${today}-\K\d{3}" | sort -n | tail -1)
|
||||
next_seq=$(printf "%03d" $(( ${last_seq:-0} + 1 )))
|
||||
```
|
||||
|
||||
Use the Write tool to save the complete plan to `docs/plans/YYYY-MM-DD-NNN-<type>-<descriptive-name>-plan.md` (where NNN is `$next_seq` from the bash command above). This step is mandatory and cannot be skipped — even when running as part of LFG/SLFG or other automated pipelines.
|
||||
|
||||
Confirm: "Plan written to docs/plans/[filename]"
|
||||
|
||||
**Pipeline mode:** If invoked from an automated workflow (LFG, SLFG, or any `disable-model-invocation` context), skip all AskUserQuestion calls. Make decisions automatically and proceed to writing the plan without interactive prompts.
|
||||
|
||||
## Output Format
|
||||
|
||||
**Filename:** Use the date and kebab-case filename from Step 2 Title & Categorization.
|
||||
**Filename:** Use the date, daily sequence number, and kebab-case filename from Step 2 Title & Categorization.
|
||||
|
||||
```
|
||||
docs/plans/YYYY-MM-DD-<type>-<descriptive-name>-plan.md
|
||||
docs/plans/YYYY-MM-DD-NNN-<type>-<descriptive-name>-plan.md
|
||||
```
|
||||
|
||||
Examples:
|
||||
- ✅ `docs/plans/2026-01-15-feat-user-authentication-flow-plan.md`
|
||||
- ✅ `docs/plans/2026-02-03-fix-checkout-race-condition-plan.md`
|
||||
- ✅ `docs/plans/2026-03-10-refactor-api-client-extraction-plan.md`
|
||||
- ❌ `docs/plans/2026-01-15-feat-thing-plan.md` (not descriptive - what "thing"?)
|
||||
- ❌ `docs/plans/2026-01-15-feat-new-feature-plan.md` (too vague - what feature?)
|
||||
- ❌ `docs/plans/2026-01-15-feat: user auth-plan.md` (invalid characters - colon and space)
|
||||
- ❌ `docs/plans/feat-user-auth-plan.md` (missing date prefix)
|
||||
- ✅ `docs/plans/2026-01-15-001-feat-user-authentication-flow-plan.md`
|
||||
- ✅ `docs/plans/2026-02-03-001-fix-checkout-race-condition-plan.md`
|
||||
- ✅ `docs/plans/2026-03-10-002-refactor-api-client-extraction-plan.md`
|
||||
- ❌ `docs/plans/2026-01-15-feat-thing-plan.md` (missing sequence number, not descriptive)
|
||||
- ❌ `docs/plans/2026-01-15-001-feat-new-feature-plan.md` (too vague - what feature?)
|
||||
- ❌ `docs/plans/2026-01-15-001-feat: user auth-plan.md` (invalid characters - colon and space)
|
||||
- ❌ `docs/plans/feat-user-auth-plan.md` (missing date prefix and sequence number)
|
||||
|
||||
## Post-Generation Options
|
||||
|
||||
After writing the plan file, use the **AskUserQuestion tool** to present these options:
|
||||
|
||||
**Question:** "Plan ready at `docs/plans/YYYY-MM-DD-<type>-<name>-plan.md`. What would you like to do next?"
|
||||
**Question:** "Plan ready at `docs/plans/YYYY-MM-DD-NNN-<type>-<name>-plan.md`. What would you like to do next?"
|
||||
|
||||
**Options:**
|
||||
1. **Open plan in editor** - Open the plan file for review
|
||||
2. **Run `/deepen-plan`** - Enhance each section with parallel research agents (best practices, performance, UI)
|
||||
3. **Run `/technical_review`** - Technical feedback from code-focused reviewers (DHH, Kieran, Simplicity)
|
||||
4. **Review and refine** - Improve the document through structured self-review
|
||||
5. **Start `/workflows:work`** - Begin implementing this plan locally
|
||||
6. **Start `/workflows:work` on remote** - Begin implementing in Claude Code on the web (use `&` to run in background)
|
||||
3. **Review and refine** - Improve the document through structured self-review
|
||||
4. **Share to Proof** - Upload to Proof for collaborative review and sharing
|
||||
5. **Start `/ce:work`** - Begin implementing this plan locally
|
||||
6. **Start `/ce:work` on remote** - Begin implementing in Claude Code on the web (use `&` to run in background)
|
||||
7. **Create Issue** - Create issue in project tracker (GitHub/Linear)
|
||||
|
||||
Based on selection:
|
||||
- **Open plan in editor** → Run `open docs/plans/<plan_filename>.md` to open the file in the user's default editor
|
||||
- **`/deepen-plan`** → Call the /deepen-plan command with the plan file path to enhance with research
|
||||
- **`/technical_review`** → Call the /technical_review command with the plan file path
|
||||
- **Review and refine** → Load `document-review` skill.
|
||||
- **`/workflows:work`** → Call the /workflows:work command with the plan file path
|
||||
- **`/workflows:work` on remote** → Run `/workflows:work docs/plans/<plan_filename>.md &` to start work in background for Claude Code web
|
||||
- **Share to Proof** → Upload the plan to Proof:
|
||||
```bash
|
||||
CONTENT=$(cat docs/plans/<plan_filename>.md)
|
||||
TITLE="Plan: <plan title from frontmatter>"
|
||||
RESPONSE=$(curl -s -X POST https://www.proofeditor.ai/share/markdown \
|
||||
-H "Content-Type: application/json" \
|
||||
-d "$(jq -n --arg title "$TITLE" --arg markdown "$CONTENT" --arg by "ai:compound" '{title: $title, markdown: $markdown, by: $by}')")
|
||||
PROOF_URL=$(echo "$RESPONSE" | jq -r '.tokenUrl')
|
||||
```
|
||||
Display: `View & collaborate in Proof: <PROOF_URL>` — skip silently if curl fails. Then return to options.
|
||||
- **`/ce:work`** → Call the /ce:work command with the plan file path
|
||||
- **`/ce:work` on remote** → Run `/ce:work docs/plans/<plan_filename>.md &` to start work in background for Claude Code web
|
||||
- **Create Issue** → See "Issue Creation" section below
|
||||
- **Other** (automatically provided) → Accept free text for rework or specific changes
|
||||
|
||||
**Note:** If running `/workflows:plan` with ultrathink enabled, automatically run `/deepen-plan` after plan creation for maximum depth and grounding.
|
||||
**Note:** If running `/ce:plan` with ultrathink enabled, automatically run `/deepen-plan` after plan creation for maximum depth and grounding.
|
||||
|
||||
Loop back to options after Simplify or Other changes until user selects `/workflows:work` or `/technical_review`.
|
||||
Loop back to options after Simplify or Other changes until user selects `/ce:work` or another action.
|
||||
|
||||
## Issue Creation
|
||||
|
||||
When user selects "Create Issue", detect their project tracker from CLAUDE.md:
|
||||
When user selects "Create Issue", detect their project tracker from AGENTS.md:
|
||||
|
||||
1. **Check for tracker preference** in user's CLAUDE.md (global or project):
|
||||
1. **Check for tracker preference** in the user's AGENTS.md (global or project). If AGENTS.md is absent, fall back to CLAUDE.md:
|
||||
- Look for `project_tracker: github` or `project_tracker: linear`
|
||||
- Or look for mentions of "GitHub Issues" or "Linear" in their workflow section
|
||||
|
||||
@@ -545,10 +633,10 @@ When user selects "Create Issue", detect their project tracker from CLAUDE.md:
|
||||
|
||||
4. **If no tracker configured:**
|
||||
Ask user: "Which project tracker do you use? (GitHub/Linear/Other)"
|
||||
- Suggest adding `project_tracker: github` or `project_tracker: linear` to their CLAUDE.md
|
||||
- Suggest adding `project_tracker: github` or `project_tracker: linear` to their AGENTS.md
|
||||
|
||||
5. **After creation:**
|
||||
- Display the issue URL
|
||||
- Ask if they want to proceed to `/workflows:work` or `/technical_review`
|
||||
- Ask if they want to proceed to `/ce:work`
|
||||
|
||||
NEVER CODE! Just research and write the plan.
|
||||
@@ -1,7 +1,7 @@
|
||||
---
|
||||
name: workflows:review
|
||||
name: ce:review
|
||||
description: Perform exhaustive code reviews using multi-agent analysis, ultra-thinking, and worktrees
|
||||
argument-hint: "[PR number, GitHub URL, branch name, or latest]"
|
||||
argument-hint: "[PR number, GitHub URL, branch name, or latest] [--serial]"
|
||||
---
|
||||
|
||||
# Review Command
|
||||
@@ -38,7 +38,7 @@ First, I need to determine the review target type and set up the code for analys
|
||||
- [ ] Determine review type: PR number (numeric), GitHub URL, file path (.md), or empty (current branch)
|
||||
- [ ] Check current git branch
|
||||
- [ ] If ALREADY on the target branch (PR branch, requested branch name, or the branch already checked out for review) → proceed with analysis on current branch
|
||||
- [ ] If DIFFERENT branch than the review target → offer to use worktree: "Use git-worktree skill for isolated Call `skill: git-worktree` with branch name
|
||||
- [ ] If DIFFERENT branch than the review target → offer to use worktree: "Use git-worktree skill for isolated Call `skill: git-worktree` with branch name"
|
||||
- [ ] Fetch PR metadata using `gh pr view --json` for title, body, files, linked issues
|
||||
- [ ] Set up language-specific analysis tools
|
||||
- [ ] Prepare security scanning environment
|
||||
@@ -53,7 +53,8 @@ Ensure that the code is ready for analysis (either in worktree or on current bra
|
||||
<protected_artifacts>
|
||||
The following paths are compound-engineering pipeline artifacts and must never be flagged for deletion, removal, or gitignore by any review agent:
|
||||
|
||||
- `docs/plans/*.md` — Plan files created by `/workflows:plan`. These are living documents that track implementation progress (checkboxes are checked off by `/workflows:work`).
|
||||
- `docs/brainstorms/*-requirements.md` — Requirements documents created by `/ce:brainstorm`. These are the product-definition artifacts that planning depends on.
|
||||
- `docs/plans/*.md` — Plan files created by `/ce:plan`. These are living documents that track implementation progress (checkboxes are checked off by `/ce:work`).
|
||||
- `docs/solutions/*.md` — Solution documents created during the pipeline.
|
||||
|
||||
If a review agent flags any file in these directories for cleanup or removal, discard that finding during synthesis. Do not create a todo for it.
|
||||
@@ -65,19 +66,52 @@ Read `compound-engineering.local.md` in the project root. If found, use `review_
|
||||
|
||||
If no settings file exists, invoke the `setup` skill to create one. Then read the newly created file and continue.
|
||||
|
||||
#### Choose Execution Mode
|
||||
|
||||
<execution_mode>
|
||||
|
||||
Before launching review agents, check for context constraints:
|
||||
|
||||
**If `--serial` flag is passed OR conversation is in a long session:**
|
||||
|
||||
Run agents ONE AT A TIME in sequence. Wait for each agent to complete before starting the next. This uses less context but takes longer.
|
||||
|
||||
**Default (parallel):**
|
||||
|
||||
Run all agents simultaneously for speed. If you hit context limits, retry with `--serial` flag.
|
||||
|
||||
**Auto-detect:** If more than 5 review agents are configured, automatically switch to serial mode and inform the user:
|
||||
"Running review agents in serial mode (6+ agents configured). Use --parallel to override."
|
||||
|
||||
</execution_mode>
|
||||
|
||||
#### Parallel Agents to review the PR:
|
||||
|
||||
<parallel_tasks>
|
||||
|
||||
**Parallel mode (default for ≤5 agents):**
|
||||
|
||||
Run all configured review agents in parallel using Task tool. For each agent in the `review_agents` list:
|
||||
|
||||
```
|
||||
Task {agent-name}(PR content + review context from settings body)
|
||||
```
|
||||
|
||||
Additionally, always run these regardless of settings:
|
||||
- Task agent-native-reviewer(PR content) - Verify new features are agent-accessible
|
||||
- Task learnings-researcher(PR content) - Search docs/solutions/ for past issues related to this PR's modules and patterns
|
||||
**Serial mode (--serial flag, or auto for 6+ agents):**
|
||||
|
||||
Run configured review agents ONE AT A TIME. For each agent in the `review_agents` list, wait for it to complete before starting the next:
|
||||
|
||||
```
|
||||
For each agent in review_agents:
|
||||
1. Task {agent-name}(PR content + review context)
|
||||
2. Wait for completion
|
||||
3. Collect findings
|
||||
4. Proceed to next agent
|
||||
```
|
||||
|
||||
Always run these last regardless of mode:
|
||||
- Task compound-engineering:review:agent-native-reviewer(PR content) - Verify new features are agent-accessible
|
||||
- Task compound-engineering:research:learnings-researcher(PR content) - Search docs/solutions/ for past issues related to this PR's modules and patterns
|
||||
|
||||
</parallel_tasks>
|
||||
|
||||
@@ -89,9 +123,9 @@ These agents are run ONLY when the PR matches specific criteria. Check the PR fi
|
||||
|
||||
**MIGRATIONS: If PR contains database migrations, schema.rb, or data backfills:**
|
||||
|
||||
- Task schema-drift-detector(PR content) - Detects unrelated schema.rb changes by cross-referencing against included migrations (run FIRST)
|
||||
- Task data-migration-expert(PR content) - Validates ID mappings match production, checks for swapped values, verifies rollback safety
|
||||
- Task deployment-verification-agent(PR content) - Creates Go/No-Go deployment checklist with SQL verification queries
|
||||
- Task compound-engineering:review:schema-drift-detector(PR content) - Detects unrelated schema.rb changes by cross-referencing against included migrations (run FIRST)
|
||||
- Task compound-engineering:review:data-migration-expert(PR content) - Validates ID mappings match production, checks for swapped values, verifies rollback safety
|
||||
- Task compound-engineering:review:deployment-verification-agent(PR content) - Creates Go/No-Go deployment checklist with SQL verification queries
|
||||
|
||||
**When to run:**
|
||||
- PR includes files matching `db/migrate/*.rb` or `db/schema.rb`
|
||||
@@ -106,7 +140,7 @@ These agents are run ONLY when the PR matches specific criteria. Check the PR fi
|
||||
|
||||
</conditional_agents>
|
||||
|
||||
### 4. Ultra-Thinking Deep Dive Phases
|
||||
### 2. Ultra-Thinking Deep Dive Phases
|
||||
|
||||
<ultrathink_instruction> For each phase below, spend maximum cognitive effort. Think step by step. Consider all angles. Question assumptions. And bring all reviews in a synthesis to the user.</ultrathink_instruction>
|
||||
|
||||
@@ -114,7 +148,7 @@ These agents are run ONLY when the PR matches specific criteria. Check the PR fi
|
||||
Complete system context map with component interactions
|
||||
</deliverable>
|
||||
|
||||
#### Phase 3: Stakeholder Perspective Analysis
|
||||
#### Phase 1: Stakeholder Perspective Analysis
|
||||
|
||||
<thinking_prompt> ULTRA-THINK: Put yourself in each stakeholder's shoes. What matters to them? What are their pain points? </thinking_prompt>
|
||||
|
||||
@@ -154,7 +188,7 @@ Complete system context map with component interactions
|
||||
- How does this affect time-to-market?
|
||||
- What's the total cost of ownership? </questions> </stakeholder_perspectives>
|
||||
|
||||
#### Phase 4: Scenario Exploration
|
||||
#### Phase 2: Scenario Exploration
|
||||
|
||||
<thinking_prompt> ULTRA-THINK: Explore edge cases and failure scenarios. What could go wrong? How does the system behave under stress? </thinking_prompt>
|
||||
|
||||
@@ -171,7 +205,7 @@ Complete system context map with component interactions
|
||||
- [ ] **Data Corruption**: Partial writes, inconsistency
|
||||
- [ ] **Cascading Failures**: Downstream service issues </scenario_checklist>
|
||||
|
||||
### 6. Multi-Angle Review Perspectives
|
||||
### 3. Multi-Angle Review Perspectives
|
||||
|
||||
#### Technical Excellence Angle
|
||||
|
||||
@@ -203,7 +237,7 @@ Complete system context map with component interactions
|
||||
|
||||
### 4. Simplification and Minimalism Review
|
||||
|
||||
Run the Task code-simplicity-reviewer() to see if we can simplify the code.
|
||||
Run the Task compound-engineering:review:code-simplicity-reviewer() to see if we can simplify the code.
|
||||
|
||||
### 5. Findings Synthesis and Todo Creation Using file-todos Skill
|
||||
|
||||
@@ -220,7 +254,7 @@ Remove duplicates, prioritize by severity and impact.
|
||||
|
||||
- [ ] Collect findings from all parallel agents
|
||||
- [ ] Surface learnings-researcher results: if past solutions are relevant, flag them as "Known Pattern" with links to docs/solutions/ files
|
||||
- [ ] Discard any findings that recommend deleting or gitignoring files in `docs/plans/` or `docs/solutions/` (see Protected Artifacts above)
|
||||
- [ ] Discard any findings that recommend deleting or gitignoring files in `docs/brainstorms/`, `docs/plans/`, or `docs/solutions/` (see Protected Artifacts above)
|
||||
- [ ] Categorize by type: security, performance, architecture, quality, etc.
|
||||
- [ ] Assign severity levels: 🔴 CRITICAL (P1), 🟡 IMPORTANT (P2), 🔵 NICE-TO-HAVE (P3)
|
||||
- [ ] Remove duplicate or overlapping findings
|
||||
@@ -401,7 +435,6 @@ After creating all todo files, present comprehensive summary:
|
||||
ls todos/*-pending-*.md # View all pending todos
|
||||
/triage # Use slash command for interactive triage
|
||||
```
|
||||
````
|
||||
|
||||
3. **Work on Approved Todos**:
|
||||
|
||||
@@ -436,10 +469,9 @@ After creating all todo files, present comprehensive summary:
|
||||
- Code cleanup
|
||||
- Optimization opportunities
|
||||
- Documentation updates
|
||||
````
|
||||
|
||||
```
|
||||
|
||||
### 7. End-to-End Testing (Optional)
|
||||
### 6. End-to-End Testing (Optional)
|
||||
|
||||
<detect_project_type>
|
||||
|
||||
@@ -525,4 +557,3 @@ The subagent will:
|
||||
### Important: P1 Findings Block Merge
|
||||
|
||||
Any **🔴 P1 (CRITICAL)** findings must be addressed before merging the PR. Present these prominently and ensure they're resolved before accepting the PR.
|
||||
```
|
||||
@@ -1,5 +1,5 @@
|
||||
---
|
||||
name: workflows:work
|
||||
name: ce:work
|
||||
description: Execute work plans efficiently while maintaining quality and finishing features
|
||||
argument-hint: "[plan file, specification, or todo file path]"
|
||||
---
|
||||
@@ -23,6 +23,10 @@ This command takes a work document (plan, specification, or todo file) and execu
|
||||
1. **Read Plan and Clarify**
|
||||
|
||||
- Read the work document completely
|
||||
- Treat the plan as a decision artifact, not an execution script
|
||||
- If the plan includes sections such as `Implementation Units`, `Work Breakdown`, `Requirements Trace`, `Files`, `Test Scenarios`, or `Verification`, use those as the primary source material for execution
|
||||
- Check for a `Deferred to Implementation` or `Implementation-Time Unknowns` section — these are questions the planner intentionally left for you to resolve during execution. Note them before starting so they inform your approach rather than surprising you mid-task
|
||||
- Check for a `Scope Boundaries` section — these are explicit non-goals. Refer back to them if implementation starts pulling you toward adjacent work
|
||||
- Review any references or links provided in the plan
|
||||
- If anything is unclear or ambiguous, ask clarifying questions now
|
||||
- Get user approval to proceed
|
||||
@@ -73,12 +77,35 @@ This command takes a work document (plan, specification, or todo file) and execu
|
||||
- You plan to switch between branches frequently
|
||||
|
||||
3. **Create Todo List**
|
||||
- Use TodoWrite to break plan into actionable tasks
|
||||
- Use your available task tracking tool (e.g., TodoWrite, task lists) to break the plan into actionable tasks
|
||||
- Derive tasks from the plan's implementation units, dependencies, files, test targets, and verification criteria
|
||||
- For each unit, read the `Patterns to follow` field before implementing — these point to specific files or conventions to mirror
|
||||
- Use each unit's `Verification` field as the primary "done" signal for that task
|
||||
- Do not expect the plan to contain implementation code, micro-step TDD instructions, or exact shell commands
|
||||
- Include dependencies between tasks
|
||||
- Prioritize based on what needs to be done first
|
||||
- Include testing and quality check tasks
|
||||
- Keep tasks specific and completable
|
||||
|
||||
4. **Choose Execution Strategy**
|
||||
|
||||
After creating the task list, decide how to execute based on the plan's size and dependency structure:
|
||||
|
||||
| Strategy | When to use |
|
||||
|----------|-------------|
|
||||
| **Inline** | 1-2 small tasks, or tasks needing user interaction mid-flight |
|
||||
| **Serial subagents** | 3+ tasks with dependencies between them. Each subagent gets a fresh context window focused on one unit — prevents context degradation across many tasks |
|
||||
| **Parallel subagents** | 3+ tasks where some units have no shared dependencies and touch non-overlapping files. Dispatch independent units simultaneously, run dependent units after their prerequisites complete |
|
||||
|
||||
**Subagent dispatch** uses your available subagent or task spawning mechanism. For each unit, give the subagent:
|
||||
- The full plan file path (for overall context)
|
||||
- The specific unit's Goal, Files, Approach, Patterns, Test scenarios, and Verification
|
||||
- Any resolved deferred questions relevant to that unit
|
||||
|
||||
After each subagent completes, update the plan checkboxes and task list before dispatching the next dependent unit.
|
||||
|
||||
For genuinely large plans needing persistent inter-agent communication (agents challenging each other's approaches, shared coordination across 10+ tasks), see Swarm Mode below which uses Agent Teams.
|
||||
|
||||
### Phase 2: Execute
|
||||
|
||||
1. **Task Execution Loop**
|
||||
@@ -87,18 +114,31 @@ This command takes a work document (plan, specification, or todo file) and execu
|
||||
|
||||
```
|
||||
while (tasks remain):
|
||||
- Mark task as in_progress in TodoWrite
|
||||
- Mark task as in-progress
|
||||
- Read any referenced files from the plan
|
||||
- Look for similar patterns in codebase
|
||||
- Implement following existing conventions
|
||||
- Write tests for new functionality
|
||||
- Run System-Wide Test Check (see below)
|
||||
- Run tests after changes
|
||||
- Mark task as completed in TodoWrite
|
||||
- Mark off the corresponding checkbox in the plan file ([ ] → [x])
|
||||
- Mark task as completed
|
||||
- Evaluate for incremental commit (see below)
|
||||
```
|
||||
|
||||
**IMPORTANT**: Always update the original plan document by checking off completed items. Use the Edit tool to change `- [ ]` to `- [x]` for each task you finish. This keeps the plan as a living document showing progress and ensures no checkboxes are left unchecked.
|
||||
**System-Wide Test Check** — Before marking a task done, pause and ask:
|
||||
|
||||
| Question | What to do |
|
||||
|----------|------------|
|
||||
| **What fires when this runs?** Callbacks, middleware, observers, event handlers — trace two levels out from your change. | Read the actual code (not docs) for callbacks on models you touch, middleware in the request chain, `after_*` hooks. |
|
||||
| **Do my tests exercise the real chain?** If every dependency is mocked, the test proves your logic works *in isolation* — it says nothing about the interaction. | Write at least one integration test that uses real objects through the full callback/middleware chain. No mocks for the layers that interact. |
|
||||
| **Can failure leave orphaned state?** If your code persists state (DB row, cache, file) before calling an external service, what happens when the service fails? Does retry create duplicates? | Trace the failure path with real objects. If state is created before the risky call, test that failure cleans up or that retry is idempotent. |
|
||||
| **What other interfaces expose this?** Mixins, DSLs, alternative entry points (Agent vs Chat vs ChatMethods). | Grep for the method/behavior in related classes. If parity is needed, add it now — not as a follow-up. |
|
||||
| **Do error strategies align across layers?** Retry middleware + application fallback + framework error handling — do they conflict or create double execution? | List the specific error classes at each layer. Verify your rescue list matches what the lower layer actually raises. |
|
||||
|
||||
**When to skip:** Leaf-node changes with no callbacks, no state persistence, no parallel interfaces. If the change is purely additive (new helper method, new view partial), the check takes 10 seconds and the answer is "nothing fires, skip."
|
||||
|
||||
**When this matters most:** Any change that touches models with callbacks, error handling with fallback/retry, or functionality exposed through multiple interfaces.
|
||||
|
||||
|
||||
2. **Incremental Commits**
|
||||
|
||||
@@ -113,6 +153,8 @@ This command takes a work document (plan, specification, or todo file) and execu
|
||||
|
||||
**Heuristic:** "Can I write a commit message that describes a complete, valuable change? If yes, commit. If the message would be 'WIP' or 'partial X', wait."
|
||||
|
||||
If the plan has Implementation Units, use them as a starting guide for commit boundaries — but adapt based on what you find during implementation. A unit might need multiple commits if it's larger than expected, or small related units might land together. Use each unit's Goal to inform the commit message.
|
||||
|
||||
**Commit workflow:**
|
||||
```bash
|
||||
# 1. Verify tests pass (use project's test command)
|
||||
@@ -134,7 +176,7 @@ This command takes a work document (plan, specification, or todo file) and execu
|
||||
- The plan should reference similar code - read those files first
|
||||
- Match naming conventions exactly
|
||||
- Reuse existing components where possible
|
||||
- Follow project coding standards (see CLAUDE.md)
|
||||
- Follow project coding standards (see AGENTS.md; use CLAUDE.md only if the repo still keeps a compatibility shim)
|
||||
- When in doubt, grep for similar implementations
|
||||
|
||||
4. **Test Continuously**
|
||||
@@ -143,8 +185,17 @@ This command takes a work document (plan, specification, or todo file) and execu
|
||||
- Don't wait until the end to test
|
||||
- Fix failures immediately
|
||||
- Add new tests for new functionality
|
||||
- **Unit tests with mocks prove logic in isolation. Integration tests with real objects prove the layers work together.** If your change touches callbacks, middleware, or error handling — you need both.
|
||||
|
||||
5. **Figma Design Sync** (if applicable)
|
||||
5. **Simplify as You Go**
|
||||
|
||||
After completing a cluster of related implementation units (or every 2-3 units), review recently changed files for simplification opportunities — consolidate duplicated patterns, extract shared helpers, and improve code reuse and efficiency. This is especially valuable when using subagents, since each agent works with isolated context and can't see patterns emerging across units.
|
||||
|
||||
Don't simplify after every single unit — early patterns may look duplicated but diverge intentionally in later units. Wait for a natural phase boundary or when you notice accumulated complexity.
|
||||
|
||||
If a `/simplify` skill or equivalent is available, use it. Otherwise, review the changed files yourself for reuse and consolidation opportunities.
|
||||
|
||||
6. **Figma Design Sync** (if applicable)
|
||||
|
||||
For UI work with Figma designs:
|
||||
|
||||
@@ -154,7 +205,7 @@ This command takes a work document (plan, specification, or todo file) and execu
|
||||
- Repeat until implementation matches design
|
||||
|
||||
6. **Track Progress**
|
||||
- Keep TodoWrite updated as you complete tasks
|
||||
- Keep the task list updated as you complete tasks
|
||||
- Note any blockers or unexpected discoveries
|
||||
- Create new tasks if scope expands
|
||||
- Keep user informed of major milestones
|
||||
@@ -169,7 +220,7 @@ This command takes a work document (plan, specification, or todo file) and execu
|
||||
# Run full test suite (use project's test command)
|
||||
# Examples: bin/rails test, npm test, pytest, go test, etc.
|
||||
|
||||
# Run linting (per CLAUDE.md)
|
||||
# Run linting (per AGENTS.md)
|
||||
# Use linting-agent before pushing to origin
|
||||
```
|
||||
|
||||
@@ -180,12 +231,14 @@ This command takes a work document (plan, specification, or todo file) and execu
|
||||
Run configured agents in parallel with Task tool. Present findings and address critical issues.
|
||||
|
||||
3. **Final Validation**
|
||||
- All TodoWrite tasks marked completed
|
||||
- All tasks marked completed
|
||||
- All tests pass
|
||||
- Linting passes
|
||||
- Code follows existing patterns
|
||||
- Figma designs match (if applicable)
|
||||
- No console errors or warnings
|
||||
- If the plan has a `Requirements Trace`, verify each requirement is satisfied by the completed work
|
||||
- If any `Deferred to Implementation` questions were noted, confirm they were resolved during execution
|
||||
|
||||
4. **Prepare Operational Validation Plan** (REQUIRED)
|
||||
- Add a `## Post-Deploy Monitoring & Validation` section to the PR description for every change.
|
||||
@@ -212,13 +265,28 @@ This command takes a work document (plan, specification, or todo file) and execu
|
||||
|
||||
Brief explanation if needed.
|
||||
|
||||
🤖 Generated with [Claude Code](https://claude.com/claude-code)
|
||||
🤖 Generated with [MODEL] via [HARNESS](HARNESS_URL) + Compound Engineering v[VERSION]
|
||||
|
||||
Co-Authored-By: Claude <noreply@anthropic.com>
|
||||
Co-Authored-By: [MODEL] ([CONTEXT] context, [THINKING]) <noreply@anthropic.com>
|
||||
EOF
|
||||
)"
|
||||
```
|
||||
|
||||
**Fill in at commit/PR time:**
|
||||
|
||||
| Placeholder | Value | Example |
|
||||
|-------------|-------|---------|
|
||||
| Placeholder | Value | Example |
|
||||
|-------------|-------|---------|
|
||||
| `[MODEL]` | Model name | Claude Opus 4.6, GPT-5.4 |
|
||||
| `[CONTEXT]` | Context window (if known) | 200K, 1M |
|
||||
| `[THINKING]` | Thinking level (if known) | extended thinking |
|
||||
| `[HARNESS]` | Tool running you | Claude Code, Codex, Gemini CLI |
|
||||
| `[HARNESS_URL]` | Link to that tool | `https://claude.com/claude-code` |
|
||||
| `[VERSION]` | `plugin.json` → `version` | 2.40.0 |
|
||||
|
||||
Subagents creating commits/PRs are equally responsible for accurate attribution.
|
||||
|
||||
2. **Capture and Upload Screenshots for UI Changes** (REQUIRED for any UI work)
|
||||
|
||||
For **any** design changes, new views, or UI modifications, you MUST capture and upload screenshots:
|
||||
@@ -292,7 +360,8 @@ This command takes a work document (plan, specification, or todo file) and execu
|
||||
|
||||
---
|
||||
|
||||
[](https://github.com/EveryInc/compound-engineering-plugin) 🤖 Generated with [Claude Code](https://claude.com/claude-code)
|
||||
[![Compound Engineering v[VERSION]](https://img.shields.io/badge/Compound_Engineering-v[VERSION]-6366f1)](https://github.com/EveryInc/compound-engineering-plugin)
|
||||
🤖 Generated with [MODEL] ([CONTEXT] context, [THINKING]) via [HARNESS](HARNESS_URL)
|
||||
EOF
|
||||
)"
|
||||
```
|
||||
@@ -312,73 +381,30 @@ This command takes a work document (plan, specification, or todo file) and execu
|
||||
|
||||
---
|
||||
|
||||
## Swarm Mode (Optional)
|
||||
## Swarm Mode with Agent Teams (Optional)
|
||||
|
||||
For complex plans with multiple independent workstreams, enable swarm mode for parallel execution with coordinated agents.
|
||||
For genuinely large plans where agents need to communicate with each other, challenge approaches, or coordinate across 10+ tasks with persistent specialized roles, use agent team capabilities if available (e.g., Agent Teams in Claude Code, multi-agent workflows in Codex).
|
||||
|
||||
### When to Use Swarm Mode
|
||||
**Agent teams are typically experimental and require opt-in.** Do not attempt to use agent teams unless the user explicitly requests swarm mode or agent teams, and the platform supports it.
|
||||
|
||||
| Use Swarm Mode when... | Use Standard Mode when... |
|
||||
|------------------------|---------------------------|
|
||||
| Plan has 5+ independent tasks | Plan is linear/sequential |
|
||||
| Multiple specialists needed (review + test + implement) | Single-focus work |
|
||||
| Want maximum parallelism | Simpler mental model preferred |
|
||||
| Large feature with clear phases | Small feature or bug fix |
|
||||
### When to Use Agent Teams vs Subagents
|
||||
|
||||
### Enabling Swarm Mode
|
||||
| Agent Teams | Subagents (standard mode) |
|
||||
|-------------|---------------------------|
|
||||
| Agents need to discuss and challenge each other's approaches | Each task is independent — only the result matters |
|
||||
| Persistent specialized roles (e.g., dedicated tester running continuously) | Workers report back and finish |
|
||||
| 10+ tasks with complex cross-cutting coordination | 3-8 tasks with clear dependency chains |
|
||||
| User explicitly requests "swarm mode" or "agent teams" | Default for most plans |
|
||||
|
||||
To trigger swarm execution, say:
|
||||
Most plans should use subagent dispatch from standard mode. Agent teams add significant token cost and coordination overhead — use them when the inter-agent communication genuinely improves the outcome.
|
||||
|
||||
> "Make a Task list and launch an army of agent swarm subagents to build the plan"
|
||||
### Agent Teams Workflow
|
||||
|
||||
Or explicitly request: "Use swarm mode for this work"
|
||||
|
||||
### Swarm Workflow
|
||||
|
||||
When swarm mode is enabled, the workflow changes:
|
||||
|
||||
1. **Create Team**
|
||||
```
|
||||
Teammate({ operation: "spawnTeam", team_name: "work-{timestamp}" })
|
||||
```
|
||||
|
||||
2. **Create Task List with Dependencies**
|
||||
- Parse plan into TaskCreate items
|
||||
- Set up blockedBy relationships for sequential dependencies
|
||||
- Independent tasks have no blockers (can run in parallel)
|
||||
|
||||
3. **Spawn Specialized Teammates**
|
||||
```
|
||||
Task({
|
||||
team_name: "work-{timestamp}",
|
||||
name: "implementer",
|
||||
subagent_type: "general-purpose",
|
||||
prompt: "Claim implementation tasks, execute, mark complete",
|
||||
run_in_background: true
|
||||
})
|
||||
|
||||
Task({
|
||||
team_name: "work-{timestamp}",
|
||||
name: "tester",
|
||||
subagent_type: "general-purpose",
|
||||
prompt: "Claim testing tasks, run tests, mark complete",
|
||||
run_in_background: true
|
||||
})
|
||||
```
|
||||
|
||||
4. **Coordinate and Monitor**
|
||||
- Team lead monitors task completion
|
||||
- Spawn additional workers as phases unblock
|
||||
- Handle plan approval if required
|
||||
|
||||
5. **Cleanup**
|
||||
```
|
||||
Teammate({ operation: "requestShutdown", target_agent_id: "implementer" })
|
||||
Teammate({ operation: "requestShutdown", target_agent_id: "tester" })
|
||||
Teammate({ operation: "cleanup" })
|
||||
```
|
||||
|
||||
See the `orchestrating-swarms` skill for detailed swarm patterns and best practices.
|
||||
1. **Create team** — use your available team creation mechanism
|
||||
2. **Create task list** — parse Implementation Units into tasks with dependency relationships
|
||||
3. **Spawn teammates** — assign specialized roles (implementer, tester, reviewer) based on the plan's needs. Give each teammate the plan file path and their specific task assignments
|
||||
4. **Coordinate** — the lead monitors task completion, reassigns work if someone gets stuck, and spawns additional workers as phases unblock
|
||||
5. **Cleanup** — shut down all teammates, then clean up the team resources
|
||||
|
||||
---
|
||||
|
||||
@@ -420,7 +446,7 @@ See the `orchestrating-swarms` skill for detailed swarm patterns and best practi
|
||||
Before creating PR, verify:
|
||||
|
||||
- [ ] All clarifying questions asked and answered
|
||||
- [ ] All TodoWrite tasks marked completed
|
||||
- [ ] All tasks marked completed
|
||||
- [ ] Tests pass (run project's test command)
|
||||
- [ ] Linting passes (use linting-agent)
|
||||
- [ ] Code follows existing patterns
|
||||
@@ -429,7 +455,7 @@ Before creating PR, verify:
|
||||
- [ ] Commit messages follow conventional format
|
||||
- [ ] PR description includes Post-Deploy Monitoring & Validation section (or explicit no-impact rationale)
|
||||
- [ ] PR description includes summary, testing notes, and screenshots
|
||||
- [ ] PR description includes Compound Engineered badge
|
||||
- [ ] PR description includes Compound Engineered badge with accurate model, harness, and version
|
||||
|
||||
## When to Use Reviewer Agents
|
||||
|
||||
@@ -449,6 +475,6 @@ For most features: tests + linting + following patterns is sufficient.
|
||||
- **Skipping clarifying questions** - Ask now, not after building wrong thing
|
||||
- **Ignoring plan references** - The plan has links for a reason
|
||||
- **Testing at the end** - Test continuously or suffer later
|
||||
- **Forgetting TodoWrite** - Track progress or lose track of what's done
|
||||
- **Forgetting to track progress** - Update task status as you go or lose track of what's done
|
||||
- **80% done syndrome** - Finish the feature, don't move on early
|
||||
- **Over-reviewing simple changes** - Save reviewer agents for complex work
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user