Skip to content

[Feature Request] Add critique skill for nine-dimension self-review of AI responses #1013

@ZhangShiZhao-Beta

Description

@ZhangShiZhao-Beta
  • I searched existing issues and this has not been proposed before

What problem does this solve?

When AI agents produce substantive responses — particularly architectural designs, technical analysis, or complex recommendations —
they often contain quality issues that aren't obvious during initial review. I've experienced this firsthand: after receiving a
complete architectural design from the AI, I asked it to confirm the design, and it found no obvious issues. But when I prompted a
structured self-examination using critical thinking standards, the AI uncovered real problems it had missed — vague abstractions,
overlooked edge cases, and logical gaps.

This is different from code review (which checks code against plans/specs) and verification-before-completion (which validates
outcomes). There's currently no skill that systematically evaluates the quality of the AI's own response across multiple
intellectual dimensions — clarity, precision, accuracy, completeness, logic, etc.

Proposed solution

A critique skill that triggers a nine-dimension self-review of the AI's most recent substantive response, based on the
well-established intellectual standards from critical thinking theory:

  1. Clarity, 2. Precision, 3. Accuracy, 4. Importance, 5. Relevance, 6. Completeness, 7. Logic, 8. Breadth, 9. Depth

Key design decisions:

  • Manual trigger only (/critique) — not automatic, to avoid slowing every interaction
  • Output only dimensions with problems found — no noise from passing dimensions
  • Severity-ordered (P0/P1/P2) with classification tags ([error], [reword], [missing], [excess])
  • Max 5 problems — keeps feedback focused and actionable
  • Language-matched — follows current conversation language
  • Waits for user direction — doesn't auto-fix, lets user choose what to address

What alternatives did you consider?

  1. Using verification-before-completion — This verifies implementation outcomes against requirements, not the quality of a
    conversational response. Different scope.
  2. Using requesting-code-review — This dispatches a subagent to review code against a plan. It doesn't evaluate the
    intellectual quality of an AI response.
  3. Simply asking "review your answer" — In practice, the AI gives superficial self-reviews without a structured framework. The
    nine-dimension approach produces significantly deeper and more actionable feedback.

Is this appropriate for core Superpowers?

Yes. This skill benefits anyone using AI coding assistants regardless of project type, language, or framework. It applies equally
to architectural discussions, debugging analysis, planning, and general conversation. It's platform-agnostic and doesn't depend on
any specific tool or domain.

Context

I've been using this as a personal skill for Claude Code with strong results. It's particularly valuable after brainstorming
sessions and architectural design discussions where the cost of undetected quality issues is high.

Metadata

Metadata

Assignees

No one assigned

    Labels

    enhancementNew feature or request

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions