windsurf-test-generationClaude Skill
Generate comprehensive test suites using Cascade.
| name | windsurf-test-generation |
| description | Generate comprehensive test suites using Cascade. Activate when users mention "generate tests", "test coverage", "write unit tests", "create test suite", or "tdd assistance". Handles AI-powered test generation. Use when writing or running tests. Trigger with phrases like "windsurf test generation", "windsurf generation", "windsurf". |
| allowed-tools | Read,Write,Edit,Bash(cmd:*),Grep,Glob |
| version | 1.0.0 |
| license | MIT |
| author | Jeremy Longshore <jeremy@intentsolutions.io> |
Windsurf Test Generation
Overview
This skill enables AI-powered test generation for any codebase using Windsurf's Cascade. It analyzes function signatures, identifies edge cases, creates meaningful assertions, and generates mock data. Supports unit tests, integration tests, and component tests across multiple frameworks including Jest, Vitest, Mocha, pytest, and testing-library.
Prerequisites
- Windsurf IDE with Cascade enabled
- Testing framework installed (Jest, Vitest, pytest, etc.)
- Project with testable code (functions, classes, components)
- Code coverage tool configured (optional but recommended)
- Understanding of testing patterns and best practices
Instructions
- Configure Testing Framework
- Select Target Code
- Generate Tests with Cascade
- Add Custom Scenarios
- Integrate into Workflow
See {baseDir}/references/implementation.md for detailed implementation guide.
Output
- Test files with comprehensive coverage
- Mock data and fixture files
- Coverage report with metrics
- Test pattern documentation for team
Error Handling
See {baseDir}/references/errors.md for comprehensive error handling.
Examples
See {baseDir}/references/examples.md for detailed examples.
Resources
Similar Claude Skills & Agent Workflows
end-to-end-tests
after making changes, run end-to-end tests to ensure that the product still works
test-coverage-improver
Improve test coverage in the OpenAI Agents Python repository: run `make coverage`, inspect coverage artifacts, identify low-coverage files, propose high-impact tests, and confirm with the user before writing tests.
code-change-verification
Run the mandatory verification stack when changes affect runtime code, tests, or build/test behavior in the OpenAI Agents Python repository.
testing-python
Write and evaluate effective Python tests using pytest.
testing
Run and troubleshoot tests for DBHub, including unit tests, integration tests with Testcontainers, and database-specific tests.
n8n-validation-expert
Interpret validation errors and guide fixing them.