Claude-code-skills ln-522-manual-tester
Performs manual testing of Story AC via executable bash scripts in tests/manual/. Use when Story implementation needs hands-on AC verification.
git clone https://github.com/levnikolaevich/claude-code-skills
T=$(mktemp -d) && git clone --depth=1 https://github.com/levnikolaevich/claude-code-skills "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills-catalog/ln-522-manual-tester" ~/.claude/skills/levnikolaevich-claude-code-skills-ln-522-manual-tester && rm -rf "$T"
skills-catalog/ln-522-manual-tester/SKILL.mdPaths: File paths (
,shared/,references/) are relative to skills repo root. If not found at CWD, locate this SKILL.md directory and go up one level for repo root. If../ln-*is missing, fetch files via WebFetch fromshared/.https://raw.githubusercontent.com/levnikolaevich/claude-code-skills/master/skills/{path}
MANDATORY READ: Load
shared/references/ci_tool_detection.md — compact output flags, pipefail, and failure-artifact policy for bash/curl/Puppeteer scripts.
Inputs
| Input | Required | Source | Description |
|---|---|---|---|
| Yes | args, git branch, kanban, user | Story to process |
Resolution: Story Resolution Chain. Status filter: To Review
Manual Tester
Type: L3 Worker
Manually verifies Story AC on running code and reports structured results for the quality gate.
Purpose & Scope
- Create executable test scripts in
folder of target project.tests/manual/ - Run AC-driven checks via bash/curl (API) or puppeteer (UI).
- Save scripts permanently for regression testing (not temp files).
- Document results in Linear with pass/fail per AC and script path.
- No status changes or task creation.
When to Use
- Use when a Story needs hands-on acceptance-criteria verification before automated planning
- Research comment "## Test Research:" exists on Story (from ln-521)
- All implementation tasks in Story status = Done
Test Design Principles
1. Fail-Fast - No Silent Failures
CRITICAL: Tests MUST return 1 (fail) immediately when any criterion is not met.
Never use:
print_status "WARN" + return 0 for validation failures, graceful degradation without explicit flags, silent fallbacks that hide errors.
Exceptions (WARN is OK): Informational warnings that don't affect correctness, optional features (with clear justification in comments), infrastructure issues (e.g., missing Nginx in dev environment).
2. Expected-Based Testing - The Golden Standard
CRITICAL: Tests MUST compare actual results against expected reference files, not apply heuristics or algorithmic checks.
Directory structure:
tests/manual/NN-feature/ ├── samples/ # Input files ├── expected/ # Expected output files (REQUIRED!) │ └── {base_name}_{source_lang}-{target_lang}.{ext} └── test-*.sh
Heuristics acceptable ONLY for: dynamic/non-deterministic data (timestamps, UUIDs, tokens - normalize before comparison; JSON with unordered keys - use
jq --sort-keys).
3. Results Storage
Test results saved to
tests/manual/results/ (persistent, in .gitignore). Named: result_{ac_name}.{ext} or response_{ac_name}.json. Inspectable after test completion for debugging.
4. Expected File Generation
To create expected files:
- Run test with current implementation
- Review output in
folderresults/ - If correct: copy to
folder with proper namingexpected/ - If incorrect: fix implementation first, then copy
IMPORTANT: Never blindly copy results to expected. Always validate correctness first.
Workflow
Phase 0: Resolve Inputs
MANDATORY READ: Load
shared/references/input_resolution_pattern.md
- Resolve storyId: Run Story Resolution Chain per guide (status filter: [To Review]).
Phase 1: Setup tests/manual structure
- Read
— get port allocation, service endpoints, base URLs. Readdocs/project/infrastructure.md
— get Docker commands, test prerequisites, environment setupdocs/project/runbook.md - Check if
folder exists in project roottests/manual/ - If missing, create structure:
— shared configuration (BASE_URL, helpers, colors)tests/manual/config.sh
— folder documentation (see README.md template below)tests/manual/README.md
— master script to run all test suites (see test-all.sh template below)tests/manual/test-all.sh
— folder for test outputs (add totests/manual/results/
).gitignore
- Add
to projecttests/manual/results/
if not present.gitignore - If exists, read existing
to reuse settings (BASE_URL, tokens)config.sh
Phase 2: Create Story test script
- Fetch Story, parse AC into Given/When/Then list (3-5 expected)
- Check for research comment (from ln-521-test-researcher) — incorporate findings into test cases
- Detect API vs UI (API → curl, UI → puppeteer). IF UI: MANDATORY READ: Load
references/puppeteer_patterns.md - Create test folder structure:
— input files (if needed)tests/manual/{NN}-{story-slug}/samples/
— expected output files (REQUIRED for deterministic tests)tests/manual/{NN}-{story-slug}/expected/
- Generate test script:
tests/manual/{NN}-{story-slug}/test-{story-slug}.sh- Use appropriate template: TEMPLATE-api-endpoint.sh (direct calls) or TEMPLATE-document-format.sh (async jobs)
- Header: Story ID, AC list, prerequisites
- Test function per AC + edge/error cases
- diff-based validation against expected files (PRIMARY)
- Results saved to
tests/manual/results/ - Summary table with timing
- Make script executable (
)chmod +x
Phase 3: Update Documentation
- Update
:tests/manual/README.md- Add new test to "Available Test Suites" table
- Include Story ID, AC covered, run command
- Update
:tests/manual/test-all.sh- Add call to new script in SUITES array
- Maintain execution order (00-setup first, then numbered suites)
Phase 4: Execute and report
MANDATORY READ: Load
references/test_result_format_v1.md
- Rebuild Docker containers (no cache), ensure healthy
- Run generated script, capture output
- Parse results (pass/fail counts)
- Post Linear comment (per test_result_format_v1.md) with:
- AC matrix (pass/fail per AC)
- Script path:
tests/manual/{NN}-{story-slug}/test-{story-slug}.sh - Rerun command:
cd tests/manual && ./{NN}-{story-slug}/test-{story-slug}.sh
Critical Rules
- Scripts saved to project
, NOT temp files.tests/manual/ - Rebuild Docker before testing; fail if rebuild/run unhealthy.
- Keep language of Story (EN/RU) in script comments and Linear comment.
- No fixes or status changes; only evidence and verdict.
- Script must be idempotent (can rerun anytime).
Runtime Summary Artifact
MANDATORY READ: Load
shared/references/test_planning_summary_contract.md, shared/references/test_planning_worker_runtime_contract.md
Runtime profile:
- family:
test-planning-worker - worker:
ln-522 - summary kind:
test-planning-worker - payload fields used by coordinators:
,worker
,status
,warningsmanual_result_path
Invocation rules:
- standalone: omit
andrunIdsummaryArtifactPath - managed: pass both
and exactrunIdsummaryArtifactPath - always write the validated summary before terminal outcome
Test scripts always go to
tests/manual/, never to the project root.
Monitor Integration (Claude Code 2.1.98+)
MANDATORY READ: Load
shared/references/monitor_integration_pattern.md
When running test scripts expected to take >30 seconds:
Monitor(command="bash tests/manual/{suite}/test-{slug}.sh 2>&1", timeout_ms=300000, description="manual test: {slug}")
Fallback: if Monitor is unavailable (Bedrock/Vertex), use
Bash(run_in_background=true).
Definition of Done
-
structure exists (config.sh, README.md, test-all.sh, results/ created if missing).tests/manual/ -
added to projecttests/manual/results/
..gitignore - Test script created at
.tests/manual/{NN}-{story-slug}/test-{story-slug}.sh -
folder created with at least 1 expected file per deterministic AC.expected/ - Script uses diff-based validation against expected files (not heuristics).
- Script saves results to
for debugging.tests/manual/results/ - Script is executable and idempotent.
- README.md updated with new test suite in "Available Test Suites" table.
- test-all.sh updated with call to new script in SUITES array.
- App rebuilt and running; tests executed.
- Verdict and Linear comment posted with script path and rerun command.
Script Templates
README.md (created once per project)
# Manual Testing Scripts > **SCOPE:** Bash scripts for manual API testing. Complements automated tests with CLI-based workflows. ## Quick Start ```bash cd tests/manual ./00-setup/create-account.sh # (if auth required) ./test-all.sh # Run ALL test suites
Prerequisites
- Docker containers running (
)docker compose ps - jq installed (
orapt-get install jq
)brew install jq
Folder Structure
tests/manual/ ├── config.sh # Shared configuration (BASE_URL, helpers, colors) ├── README.md # This file ├── test-all.sh # Run all test suites ├── 00-setup/ # Account & token setup (if auth required) │ ├── create-account.sh │ └── get-token.sh └── {NN}-{topic}/ # Test suites by Story └── test-{slug}.sh
Available Test Suites
<!-- Add new test suites here when creating new tests -->| Suite | Story | AC Covered | Run Command |
|---|---|---|---|
| — | — | — | — |
Adding New Tests
- Create script in
{NN}-{topic}/test-{slug}.sh - Update this README (Available Test Suites table)
- Update
(add to SUITES array)test-all.sh
### test-all.sh (created once per project) ```bash #!/bin/bash # ============================================================================= # Run all manual test suites # ============================================================================= set -e SCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)" source "$SCRIPT_DIR/config.sh" echo "==========================================" echo "Running ALL Manual Test Suites" echo "==========================================" check_jq check_api # Setup (if exists) [ -f "$SCRIPT_DIR/00-setup/create-account.sh" ] && "$SCRIPT_DIR/00-setup/create-account.sh" [ -f "$SCRIPT_DIR/00-setup/get-token.sh" ] && "$SCRIPT_DIR/00-setup/get-token.sh" # Test suites (add new suites here) SUITES=( # "01-auth/test-auth-flow.sh" # "02-translation/test-translation.sh" ) PASSED=0; FAILED=0 for suite in "${SUITES[@]}"; do echo "" echo "==========================================" echo "Running: $suite" echo "==========================================" if "$SCRIPT_DIR/$suite"; then ((++PASSED)) print_status "PASS" "$suite" else ((++FAILED)) print_status "FAIL" "$suite" fi done echo "" echo "==========================================" echo "TOTAL: $PASSED suites passed, $FAILED failed" echo "==========================================" [ $FAILED -eq 0 ] && exit 0 || exit 1
config.sh (created once per project)
#!/bin/bash # Shared configuration for manual testing scripts export BASE_URL="${BASE_URL:-http://localhost:8080}" export RED='\033[0;31m' export GREEN='\033[0;32m' export YELLOW='\033[1;33m' export NC='\033[0m' print_status() { local status=$1; local message=$2 case $status in "PASS") echo -e "${GREEN}[PASS]${NC} $message" ;; "FAIL") echo -e "${RED}[FAIL]${NC} $message" ;; "WARN") echo -e "${YELLOW}[WARN]${NC} $message" ;; "INFO") echo -e "[INFO] $message" ;; esac } check_jq() { command -v jq &> /dev/null || { echo "Error: jq required"; exit 1; } } check_api() { local response=$(curl -s -o /dev/null -w "%{http_code}" "$BASE_URL/health" 2>/dev/null) if [ "$response" != "200" ]; then echo "Error: API not reachable at $BASE_URL" exit 1 fi print_status "INFO" "API reachable at $BASE_URL" } SCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)" export SCRIPT_DIR
Test Script Templates
| Template | Use Case | Location |
|---|---|---|
| template-api-endpoint.sh | API endpoint tests (NO async jobs) | template-api-endpoint.sh |
| template-document-format.sh | Document/file processing (WITH async jobs) | template-document-format.sh |
Quick start:
cp references/templates/template-api-endpoint.sh {NN}-feature/test-{feature}.sh # Endpoint tests cp references/templates/template-document-format.sh {NN}-feature/test-{format}.sh # Document tests
Reference Files
- Script format reference: prompsit-api
(production example)tests/manual/ - AC format:
(or localshared/templates/test_task_template.md
in target project)docs/templates/ - Risk-based context:
shared/references/risk_based_testing_guide.md - Research findings: ln-521-test-researcher creates "## Test Research" comment on Story
- Puppeteer patterns:
references/puppeteer_patterns.md - Test result format:
references/test_result_format_v1.md
Version: 1.0.0 Last Updated: 2026-01-15