Agent-almanac write-validation-documentation
git clone https://github.com/pjt222/agent-almanac
T=$(mktemp -d) && git clone --depth=1 https://github.com/pjt222/agent-almanac "$T" && mkdir -p ~/.claude/skills && cp -r "$T/i18n/caveman/skills/write-validation-documentation" ~/.claude/skills/pjt222-agent-almanac-write-validation-documentation-c94195 && rm -rf "$T"
i18n/caveman/skills/write-validation-documentation/SKILL.mdWrite Validation Documentation
Create complete IQ/OQ/PQ validation documentation for computerized systems.
When to Use
- Validating R or other software for regulated use
- Preparing for regulatory audit
- Documenting qualification of computing environments
- Creating or updating validation protocols and reports
Inputs
- Required: System/software to validate (name, version, purpose)
- Required: Validation plan defining scope and strategy
- Required: User requirements specification
- Optional: Existing SOP templates
- Optional: Previous validation documentation (for re-qualification)
Procedure
Step 1: Write Installation Qualification (IQ) Protocol
# Installation Qualification Protocol **System**: R Statistical Computing Environment **Version**: 4.5.0 **Document ID**: IQ-PROJ-001 **Prepared by**: [Name] | **Date**: [Date] **Reviewed by**: [Name] | **Date**: [Date] **Approved by**: [Name] | **Date**: [Date] ## 1. Objective Verify that R and required packages are correctly installed per specifications. ## 2. Prerequisites - [ ] Server/workstation meets hardware requirements - [ ] Operating system qualified - [ ] Network access available (for package downloads) ## 3. Test Cases ### IQ-001: R Installation | Field | Value | |-------|-------| | Requirement | R version 4.5.0 correctly installed | | Procedure | Open R console, execute `R.version.string` | | Expected Result | "R version 4.5.0 (2025-04-11)" | | Actual Result | ______________________ | | Pass/Fail | [ ] | | Executed by | ____________ Date: ________ | ### IQ-002: Package Inventory | Package | Required Version | Installed Version | Pass/Fail | |---------|-----------------|-------------------|-----------| | dplyr | 1.1.4 | | [ ] | | ggplot2 | 3.5.0 | | [ ] | | survival | 3.7-0 | | [ ] | ## 4. Deviations [Document any deviations from expected results and their resolution] ## 5. Conclusion [ ] All IQ tests PASSED - system installation verified [ ] IQ tests FAILED - see deviation section
Expected:
validation/iq/iq_protocol.md is complete with a unique document ID, objective, prerequisites checklist, test cases for R installation and every required package, deviation section, and approval fields.
On failure: If the organization requires a different document format, adapt the template to match the existing SOP. The key fields (requirement, procedure, expected result, actual result, pass/fail) must be preserved regardless of format.
Step 2: Write Operational Qualification (OQ) Protocol
# Operational Qualification Protocol **Document ID**: OQ-PROJ-001 ## 1. Objective Verify that the system operates correctly under normal conditions. ## 2. Test Cases ### OQ-001: Data Import Functionality | Field | Value | |-------|-------| | Requirement | System correctly imports CSV files | | Test Data | validation/test_data/import_test.csv (MD5: abc123) | | Procedure | Execute `read.csv("import_test.csv")` | | Expected | Data frame with 100 rows, 5 columns | | Actual Result | ______________________ | | Evidence | Screenshot/log file reference | ### OQ-002: Statistical Calculations | Field | Value | |-------|-------| | Requirement | t-test produces correct results | | Test Data | Known dataset: x = c(2.1, 2.5, 2.3), y = c(3.1, 3.5, 3.3) | | Procedure | Execute `t.test(x, y)` | | Expected | t = -5.000, df = 4, p = 0.00753 | | Actual Result | ______________________ | | Tolerance | ±0.001 | ### OQ-003: Error Handling | Field | Value | |-------|-------| | Requirement | System handles invalid input gracefully | | Procedure | Execute `analysis_function(invalid_input)` | | Expected | Informative error message, no crash | | Actual Result | ______________________ |
Expected:
validation/oq/oq_protocol.md contains test cases for data import, statistical calculations, and error handling, each with specific test data, expected results (with tolerances where applicable), and evidence requirements.
On failure: If test data is not yet available, create synthetic test datasets with known properties. Document the data generation method so results can be independently verified.
Step 3: Write Performance Qualification (PQ) Protocol
# Performance Qualification Protocol **Document ID**: PQ-PROJ-001 ## 1. Objective Verify the system performs as intended with real-world data and workflows. ## 2. Test Cases ### PQ-001: End-to-End Primary Analysis | Field | Value | |-------|-------| | Requirement | Primary endpoint analysis matches reference | | Test Data | Blinded test dataset (hash: sha256:abc...) | | Reference | Independent SAS calculation (report ref: SAS-001) | | Procedure | Execute full analysis pipeline | | Expected | Estimate within ±0.001 of reference | | Actual Result | ______________________ | ### PQ-002: Report Generation | Field | Value | |-------|-------| | Requirement | Generated report contains all required sections | | Procedure | Execute report generation script | | Checklist | | | | [ ] Title page with study information | | | [ ] Table of contents | | | [ ] Demographic summary table | | | [ ] Primary analysis results | | | [ ] Appendix with session info |
Expected:
validation/pq/pq_protocol.md contains end-to-end test cases using real-world (or representative) data, with results compared against an independent reference calculation (e.g., SAS output). Tolerances are explicitly defined.
On failure: If independent reference results are not available, document the gap and use dual-programming (two independent R implementations) as an alternative verification method. Flag the PQ as provisional until independent verification is complete.
Step 4: Write Qualification Reports
After executing protocols, document results:
# Installation Qualification Report **Document ID**: IQ-RPT-001 **Protocol Reference**: IQ-PROJ-001 ## 1. Summary All IQ test cases were executed on [date] by [name]. ## 2. Results Summary | Test ID | Description | Result | |---------|-------------|--------| | IQ-001 | R Installation | PASS | | IQ-002 | Package Inventory | PASS | ## 3. Deviations None observed. ## 4. Conclusion The installation of R 4.5.0 and associated packages has been verified and meets all specified requirements. ## 5. Approvals | Role | Name | Signature | Date | |------|------|-----------|------| | Executor | | | | | Reviewer | | | | | Approver | | | |
Expected: Qualification reports (IQ, OQ, PQ) are complete with all test results filled in, deviations documented (or "None observed"), conclusions stated, and approval signature fields ready for sign-off.
On failure: If test failures occurred during execution, document each failure as a deviation with root cause analysis and resolution. Do not leave deviation sections blank when failures were observed.
Step 5: Automate Where Possible
Create automated test scripts that generate evidence:
# validation/scripts/run_iq.R sink("validation/iq/iq_evidence.txt") cat("IQ Execution Date:", format(Sys.time()), "\n\n") cat("IQ-001: R Version\n") cat("Result:", R.version.string, "\n") cat("Status:", ifelse(R.version$major == "4" && R.version$minor == "5.0", "PASS", "FAIL"), "\n\n") cat("IQ-002: Package Versions\n") required <- renv::dependencies() installed <- installed.packages() # ... comparison logic sink()
Expected: Automated scripts in
validation/scripts/ generate evidence files (e.g., iq_evidence.txt) with timestamped results for each test case, reducing manual data entry and ensuring reproducibility.
On failure: If automated scripts fail due to environment differences, run them manually and capture output with
sink(). Document any differences between automated and manual execution in the qualification report.
Validation
- All protocols have unique document IDs
- Protocols reference the validation plan
- Test cases have clear pass/fail criteria
- Reports include all executed test results
- Deviations are documented with resolutions
- Approval signatures are obtained
- Documents follow organization's SOP templates
Common Pitfalls
- Vague acceptance criteria: "System works correctly" is not testable. Specify exact expected values.
- Missing evidence: Every test result needs supporting evidence (screenshots, logs, output files)
- Incomplete deviation handling: All failures must be documented, investigated, and resolved
- No version control for documents: Validation docs need change control just like code
- Skipping re-qualification: System updates (R version, package updates) require re-qualification assessment
Related Skills
- project structure for validated environmentssetup-gxp-r-project
- electronic records trackingimplement-audit-trail
- output validation methodologyvalidate-statistical-output