Generate comprehensive test reports with metrics, coverage, and visualizations. Use when performing specialized testing. Trigger with phrases like "generate test report", "create test documentation", or "show test metrics".
Installation
Details
Usage
After installing, this skill will be available to your AI coding assistant.
Verify installation:
npx agent-skills-cli listSkill Instructions
name: generating-test-reports description: | Generate comprehensive test reports with metrics, coverage, and visualizations. Use when performing specialized testing. Trigger with phrases like "generate test report", "create test documentation", or "show test metrics".
allowed-tools: Read, Write, Edit, Grep, Glob, Bash(test:report-*) version: 1.0.0 author: Jeremy Longshore jeremy@intentsolutions.io license: MIT
Test Report Generator
This skill provides automated assistance for test report generator tasks.
Prerequisites
Before using this skill, ensure you have:
- Test environment configured and accessible
- Required testing tools and frameworks installed
- Test data and fixtures prepared
- Appropriate permissions for test execution
- Network connectivity if testing external services
Instructions
Step 1: Prepare Test Environment
Set up the testing context:
- Use Read tool to examine configuration from {baseDir}/config/
- Validate test prerequisites are met
- Initialize test framework and load dependencies
- Configure test parameters and thresholds
Step 2: Execute Tests
Run the test suite:
- Use Bash(test:report-*) to invoke test framework
- Monitor test execution progress
- Capture test outputs and metrics
- Handle test failures and error conditions
Step 3: Analyze Results
Process test outcomes:
- Identify passed and failed tests
- Calculate success rate and performance metrics
- Detect patterns in failures
- Generate insights for improvement
Step 4: Generate Report
Document findings in {baseDir}/test-reports/:
- Test execution summary
- Detailed failure analysis
- Performance benchmarks
- Recommendations for fixes
Output
The skill generates comprehensive test results:
Test Summary
- Total tests executed
- Pass/fail counts and percentage
- Execution time metrics
- Resource utilization stats
Detailed Results
Each test includes:
- Test name and identifier
- Execution status (pass/fail/skip)
- Actual vs. expected outcomes
- Error messages and stack traces
Metrics and Analysis
- Code coverage percentages
- Performance benchmarks
- Trend analysis across runs
- Quality gate compliance status
Error Handling
Common issues and solutions:
Environment Setup Failures
- Error: Test environment not properly configured
- Solution: Verify configuration files; check environment variables; ensure dependencies are installed
Test Execution Timeouts
- Error: Tests exceeded maximum execution time
- Solution: Increase timeout thresholds; optimize slow tests; parallelize test execution
Resource Exhaustion
- Error: Insufficient memory or disk space during testing
- Solution: Clean up temporary files; reduce concurrent test workers; increase resource allocation
Dependency Issues
- Error: Required services or databases unavailable
- Solution: Verify service health; check network connectivity; use mocks if services are down
Resources
Testing Tools
- Industry-standard testing frameworks for your language/platform
- CI/CD integration guides and plugins
- Test automation best practices documentation
Best Practices
- Maintain test isolation and independence
- Use meaningful test names and descriptions
- Keep tests fast and focused
- Implement proper setup and teardown
- Version control test artifacts
- Run tests in CI/CD pipelines
Overview
This skill provides automated assistance for test report generator tasks. This skill provides automated assistance for the described functionality.
Examples
Example usage patterns will be demonstrated in context.
More by HelixDevelopment
View allManage master Cursor AI chat interface for code assistance. Triggers on "cursor chat", "cursor ai chat", "ask cursor", "cursor conversation", "chat with cursor". Use when working with cursor ai chat functionality. Trigger with phrases like "cursor ai chat", "cursor chat", "cursor".
Validate load balancer behavior, failover, and traffic distribution. Use when performing specialized testing. Trigger with phrases like "test load balancer", "validate failover", or "check traffic distribution".
Monitor Granola usage, analytics, and meeting insights. Use when tracking meeting patterns, analyzing team productivity, or building meeting analytics dashboards. Trigger with phrases like "granola analytics", "granola metrics", "granola monitoring", "meeting insights", "granola observability".
Validate API responses against schemas to ensure contract compliance and data integrity. Use when ensuring API response correctness. Trigger with phrases like "validate responses", "check API responses", or "verify response format".
