Analyze whether your response addresses the actual question asked before posting. Use when: (1) About to post response to forum/Slack question, (2) Want to validate response coverage, (3) Need to ensure solving the right problem, (4) Want specific improvement suggestions for gaps in response
Installation
Details
Usage
After installing, this skill will be available to your AI coding assistant.
Verify installation:
npx agent-skills-cli listSkill Instructions
name: response-quality-analysis description: "Analyze whether your response addresses the actual question asked before posting. Use when: (1) About to post response to forum/Slack question, (2) Want to validate response coverage, (3) Need to ensure solving the right problem, (4) Want specific improvement suggestions for gaps in response" license: Public domain - shared for AI engineering community
Response Quality Analysis
Validates that your response actually solves the problem asked, not the problem you're comfortable addressing. Systematically analyzes problem quality, decomposes into components, calculates coverage, and provides actionable improvements.
When to Use
Before posting substantive answers to:
- Forum questions (Slack, internal channels, Stack Overflow)
- Mailing list responses
- Documentation contributions
- Any situation where you want to ensure helpfulness
Quick Start
Ask user for:
original_problem- The question/problem statementdraft_response- Your proposed answerresponse_context- Where posting (optional)work_dir- Artifact location (default: ".sop/response-analysis")
Support multiple input methods: direct paste, screenshot, link, file.
Workflow
1. Problem Intake Validation (Score: X/5)
Check original problem for required elements:
- β Specific question (concrete, not "help me with X")
- β Current state (what exists now)
- β Desired state (what "solved" looks like)
- β οΈ Constraints (environment, tools, culture)
- β Prior attempts (what's been tried)
Scoring:
- β Present (clearly stated)
- β οΈ Partial (mentioned but vague)
- β Missing (not addressed)
Save to {work_dir}/intake-validation.md
Present to user:
Problem Quality: X/5 elements
β
Present: [list with evidence]
β οΈ Partial: [list with what's unclear]
β Missing: [list with why needed]
Options:
[A] Proceed (score >= 3/5)
[B] Generate clarifying questions for OP
[C] Note limitations, proceed anyway
[D] Too vague - cannot analyze
MUST NOT proceed if < 3/5 without user confirmation
2. Problem Decomposition (N components)
Extract components ONLY from evidence in original problem.
Format each:
## Component: [Name]
**Evidence:** "[Direct quote]"
**What they're asking:** [Specific need]
**Success criteria:** [Observable outcome]
**Priority:** Critical / Important / Nice-to-have
Priority based on:
- Critical: Core to question, must address
- Important: Explicitly mentioned or strongly implied
- Nice-to-have: Would add value, not strictly necessary
Target: 2-5 components (more = not distinct)
Save to {work_dir}/decomposition.md
Checkpoint:
I've identified [N] components:
1. [Name] - [Priority] - [One line]
2. [Name] - [Priority] - [One line]
Questions:
- Match your understanding?
- Missed or incorrect components?
- Adjust priorities?
MUST get user confirmation before proceeding
3. Coverage Analysis (Overall: Y%)
Map draft response to each component.
Coverage calculation:
- 0% - Not mentioned
- 1-40% - Mentioned but vague (no specifics)
- 41-80% - Addressed with details but incomplete
- 81-100% - Fully addressed, concrete, actionable
For each component:
## Component: [Name] - Coverage: X%
**Needed:** [Success criteria]
**Provided:** "[Quote from response]"
**Analysis:**
β
Strengths: [What covered well]
β Gaps: [What missing/vague]
π Specificity: [Concrete enough?]
**Justification:** [Why this score]
Calculate weighted coverage:
- Critical: 2x weight
- Important: 1x weight
- Nice-to-have: 0.5x weight
Coverage Matrix:
Component | Priority | Coverage | Weight | Contribution
--------------------------------------------------------
[Name] | Critical | X% | 2.0 | Y
[Name] | Important | X% | 1.0 | Y
Save to {work_dir}/coverage-analysis.md
MUST NOT inflate scores - be honest
4. Context Grounding Check
Verify response fits original context.
Check for:
-
Environment mismatch:
- Different org structure?
- Requires tools/processes not mentioned?
- Assumes different culture?
-
Problem substitution:
- Solving YOUR past problem?
- Related but different problem?
- More general problem?
-
Constraint violation:
- Violates stated constraints?
- Requires unavailable resources?
- Assumes capabilities they lack?
Document each:
## Mismatch: [Type]
**They have:** [Their context]
**You assumed:** [Your assumption]
**Impact:** [Why matters]
**Severity:** Minor / Moderate / Critical
Summary:
Context Grounding: Strong / Moderate / Weak
- Environment: Matches / Partial / Mismatches
- Problem: Same / Related / Different
- Constraints: Respected / Some violations / Major violations
Save to {work_dir}/context-check.md
Ask:
Are you solving THEIR problem or similar one you faced?
[A] Their exact problem
[B] Similar (need to adapt)
[C] Different (reconsider approach)
5. Response Quality Validation (Z/6 tests)
Run testable conditions:
Test 1: Restatement (Pass/Fail)
- Can you restate problem accurately?
- Would OP say "yes, that's it"?
Test 2: Coverage (Pass/Partial/Fail)
- Weighted coverage >= 80%? (Pass)
- All Critical >= 80%? (Pass)
-
= 60% OR Criticals >= 60%? (Partial)
Test 3: Specificity (Pass/Partial/Fail)
- Can implement tomorrow?
- Includes concrete mechanisms?
Test 4: Gap Acknowledgment (Pass/Fail)
- Explicitly stated what not addressed?
Test 5: Context Grounding (Pass/Partial/Fail)
- Fits their environment?
- No critical mismatches?
Test 6: Implementation Path (Pass/Partial/Fail)
- Can trace: current β advice β solved?
- Steps connected without gaps?
Scoring:
- Pass = 1.0
- Partial = 0.5
- Fail = 0.0
Present:
Response Quality Assessment
===========================
Coverage: X% (weighted)
Quality Score: Y/6 tests
β
PASS: [list]
β οΈ PARTIAL: [list]
β FAIL: [list]
Overall: Excellent / Good / Needs Work / Poor
Critical Gap: [Most important missing piece]
Recommendation: [Specific action]
Save to {work_dir}/validation-results.md
MUST be honest - don't inflate
6. Improvement Suggestions
For each gap (coverage < 80%):
## Gap: [Component] - Currently X% coverage
**What's missing:** [Specific elements needed]
**Current text:** "[Quote or indicate missing]"
**Suggested addition:**
"[Concrete text to add - be specific enough to copy/paste]"
**Why helps:** [How addresses gap]
**Priority:** Critical / Important / Enhancement
Order by priority (Critical first)
MUST NOT provide vague suggestions:
β "Add more detail"
β
"Add: 'For intake, create shared spreadsheet where Teams A,B,C submit requests with [fields]...'"
Present:
Improvement Suggestions
======================
I've identified [N] gaps:
Critical (must fix):
1. [Component]: [Gap summary]
Important (should fix):
2. [Component]: [Gap summary]
Which gaps address?
[A] All critical
[B] Critical + important
[C] Specific gaps (which?)
[D] None - proceed to decision
Save to {work_dir}/improvements.md
7. Revision Support (if requested)
Work on ONE gap at a time.
For each:
Revision: [Component]
=====================
Coverage: X% β Target: Y%
--- BEFORE ---
[Current text or "Not addressed"]
--- SUGGESTED ADDITION ---
[New text]
--- INTEGRATED RESULT ---
[How fits into full response]
New coverage: Y%
Approve?
[A] Yes, apply
[B] Modify (how?)
[C] Skip this gap
MUST:
- Wait for approval on each
- Update
{work_dir}/draft-response.md - Track in
{work_dir}/revision-history.md - Re-run coverage analysis after all revisions
Show impact:
Revision Impact
===============
Before: X% coverage, Y/6 quality
After: X% coverage, Y/6 quality
Improved:
- [Component]: X% β Y%
- [Test]: FAIL β PASS
8. Decision Point
Present final assessment:
Final Response Assessment
=========================
Problem Quality: X/5
Response Coverage: Y%
Quality Score: Z/6
Overall: [Excellent/Good/Acceptable/Needs Work/Poor]
Strengths:
- [What covered well]
Remaining Gaps:
- [Critical/important gaps]
Context Alignment: [Strong/Adequate/Weak]
Recommendation: [Specific advice]
Options:
[A] Post as-is
β Coverage adequate
β Acknowledge limitations
[B] Make revisions
β Address: [gaps]
β Expected: X% β Y%
[C] Reconsider approach
β May not fit context
β Consider: [alternative]
[D] Request clarification first
β Missing: [what need]
β Generate questions
[E] Don't post
β Coverage too low
β Start over
MUST NOT recommend posting if:
- Critical components < 50% coverage
- Severe context mismatches
- Quality score < 3/6
Save to {work_dir}/final-assessment.md
9. Post-Response Support (optional)
If posting, offer help with:
[A] Gap acknowledgment section:
"This addresses [covered], but doesn't cover [gaps].
For [gap], consider [pointer]."
[B] Frame limitations appropriately
- Don't over-apologize
- Frame as "areas for discussion"
[C] Add clarifying questions to OP
[D] Format for platform (Slack/forum)
Save final to {work_dir}/final-response.md
10. Learning Capture (optional)
Document insights in {work_dir}/learnings.md:
## Common Issue: [Pattern]
Example: "I solve feature overlap vs dependency coordination"
## Improvement
Watch for: [What to check next time]
## Meta-patterns
[Recurring blind spots]
Offer: "Analyze past responses for patterns?"
Example: Slack Response
Problem (3/5 quality): "How to handle multiple xfn dependency requests? Work flying everywhere (left diagram). Want centralized coordination (right diagram)."
Components identified:
- Intake mechanism (Critical) - 40% coverage
- Prioritization framework (Critical) - 0% coverage
- Communication (Important) - 30% coverage
- Transition path (Important) - 10% coverage
Initial coverage: 16% weighted
Critical gap: No prioritization method
After revisions: 68% coverage
Key improvement: Added concrete prioritization mechanism: "Establish weekly dependency review where TPM scores requests by impact/urgency/capacity, publishes priority queue visible to all requesters."
Troubleshooting
Low coverage despite effort:
- Solving related but different problem
- Go back to decomposition - reread question
- Map each paragraph to components
- Cut unmapped content
Context mismatches:
- List all assumptions made
- Check each against original
- Use conditional language: "If you have X..."
Tests pass but feels wrong:
- Being too generous with scores
- Re-read original fresh
- Ask: "Would this satisfy me?"
Takes too long:
- Simple questions: Skip to Step 3
- Obvious gaps: Skip validation
- Use judgment on rigor needed
Quick Heuristics
- Simple factual Q β Don't need full process
- Advice/approach Q β Use full process
- Vague/complex Q β Definitely use full process
Key Constraints
- MUST ask for both original_problem and draft_response upfront
- MUST NOT proceed with analysis if problem score < 3/5 without confirmation
- MUST NOT add components without evidence in original text
- MUST NOT inflate coverage scores - be honest
- MUST wait for user confirmation at checkpoints
- MUST work on one revision at a time with approval
- MUST NOT recommend posting if critical components < 50% or quality < 3/6
Meta-Validation
Before completing, verify:
- All 5 phases executed
- User confirmed at checkpoints
- Coverage honestly assessed
- Context mismatches identified
- Specific improvements provided
- Final recommendation justified
Version
v1.0 - Based on universal AI workflow pattern, tested with real Slack example
More by m31uk3
View allAnalyze a codebase and generate comprehensive documentation including architecture, components, interfaces, workflows, and dependencies. Creates an AI-optimized knowledge base (index.md) and can consolidate into AGENTS.md, README.md, or CONTRIBUTING.md. Use when the user wants to document a codebase, create AGENTS.md, understand system architecture, generate developer documentation, or asks to "summarize the codebase".
Guide for creating reliable AI workflows and SOPs. Use when: (1) User wants to create a structured workflow for AI tasks, (2) User needs to build an SOP for complex processes, (3) User wants to ensure their workflow follows best practices for managing LLM uncertainty, (4) User mentions creating workflows for domains like code review, response analysis, documentation, or any structured process
This skill should be used when the user asks to "add resiliency to a skill", "make this skill more robust", "improve error handling", "add validation mechanisms", "create self-correcting behavior", or discusses determinism, robustness, error correction, or homeostatic patterns in Agent Skills. Applies biological resiliency principles from Michael Levin's work to Agent Skill design.
Transform rough ideas into detailed design documents with implementation plans. Use when a user wants to develop an idea into a complete specification, create a design document from a concept, plan a feature implementation, or mentions "PDD", "prompt-driven development", "idea to design", "design doc from idea", or wants to systematically refine requirements before building. Guides through requirements clarification, research, detailed design, and implementation planning.
