What problem does this solve?
When AI agents produce substantive responses — particularly architectural designs, technical analysis, or complex recommendations —
they often contain quality issues that aren't obvious during initial review. I've experienced this firsthand: after receiving a
complete architectural design from the AI, I asked it to confirm the design, and it found no obvious issues. But when I prompted a
structured self-examination using critical thinking standards, the AI uncovered real problems it had missed — vague abstractions,
overlooked edge cases, and logical gaps.
This is different from code review (which checks code against plans/specs) and verification-before-completion (which validates
outcomes). There's currently no skill that systematically evaluates the quality of the AI's own response across multiple
intellectual dimensions — clarity, precision, accuracy, completeness, logic, etc.
Proposed solution
A critique skill that triggers a nine-dimension self-review of the AI's most recent substantive response, based on the
well-established intellectual standards from critical thinking theory:
- Clarity, 2. Precision, 3. Accuracy, 4. Importance, 5. Relevance, 6. Completeness, 7. Logic, 8. Breadth, 9. Depth
Key design decisions:
- Manual trigger only (
/critique) — not automatic, to avoid slowing every interaction
- Output only dimensions with problems found — no noise from passing dimensions
- Severity-ordered (P0/P1/P2) with classification tags (
[error], [reword], [missing], [excess])
- Max 5 problems — keeps feedback focused and actionable
- Language-matched — follows current conversation language
- Waits for user direction — doesn't auto-fix, lets user choose what to address
What alternatives did you consider?
- Using
verification-before-completion — This verifies implementation outcomes against requirements, not the quality of a
conversational response. Different scope.
- Using
requesting-code-review — This dispatches a subagent to review code against a plan. It doesn't evaluate the
intellectual quality of an AI response.
- Simply asking "review your answer" — In practice, the AI gives superficial self-reviews without a structured framework. The
nine-dimension approach produces significantly deeper and more actionable feedback.
Is this appropriate for core Superpowers?
Yes. This skill benefits anyone using AI coding assistants regardless of project type, language, or framework. It applies equally
to architectural discussions, debugging analysis, planning, and general conversation. It's platform-agnostic and doesn't depend on
any specific tool or domain.
Context
I've been using this as a personal skill for Claude Code with strong results. It's particularly valuable after brainstorming
sessions and architectural design discussions where the cost of undetected quality issues is high.
What problem does this solve?
When AI agents produce substantive responses — particularly architectural designs, technical analysis, or complex recommendations —
they often contain quality issues that aren't obvious during initial review. I've experienced this firsthand: after receiving a
complete architectural design from the AI, I asked it to confirm the design, and it found no obvious issues. But when I prompted a
structured self-examination using critical thinking standards, the AI uncovered real problems it had missed — vague abstractions,
overlooked edge cases, and logical gaps.
This is different from code review (which checks code against plans/specs) and verification-before-completion (which validates
outcomes). There's currently no skill that systematically evaluates the quality of the AI's own response across multiple
intellectual dimensions — clarity, precision, accuracy, completeness, logic, etc.
Proposed solution
A
critiqueskill that triggers a nine-dimension self-review of the AI's most recent substantive response, based on thewell-established intellectual standards from critical thinking theory:
Key design decisions:
/critique) — not automatic, to avoid slowing every interaction[error],[reword],[missing],[excess])What alternatives did you consider?
verification-before-completion— This verifies implementation outcomes against requirements, not the quality of aconversational response. Different scope.
requesting-code-review— This dispatches a subagent to review code against a plan. It doesn't evaluate theintellectual quality of an AI response.
nine-dimension approach produces significantly deeper and more actionable feedback.
Is this appropriate for core Superpowers?
Yes. This skill benefits anyone using AI coding assistants regardless of project type, language, or framework. It applies equally
to architectural discussions, debugging analysis, planning, and general conversation. It's platform-agnostic and doesn't depend on
any specific tool or domain.
Context
I've been using this as a personal skill for Claude Code with strong results. It's particularly valuable after brainstorming
sessions and architectural design discussions where the cost of undetected quality issues is high.