install
source · Clone the upstream repo
git clone https://github.com/mdbabumiamssm/LLMs-Universal-Life-Science-and-Clinical-Skills-
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/mdbabumiamssm/LLMs-Universal-Life-Science-and-Clinical-Skills- "$T" && mkdir -p ~/.claude/skills && cp -r "$T/Skills/Clinical/Safety/AI_Safety_Auditor" ~/.claude/skills/mdbabumiamssm-llms-universal-life-science-and-clinical-skills-ai-safety-auditor && rm -rf "$T"
manifest:
Skills/Clinical/Safety/AI_Safety_Auditor/SKILL.mdsource content
<!--
# COPYRIGHT NOTICE
# This file is part of the "Universal Biomedical Skills" project.
# Copyright (c) 2026 MD BABU MIA, PhD <md.babu.mia@mssm.edu>
# All Rights Reserved.
#
# This code is proprietary and confidential.
# Unauthorized copying of this file, via any medium is strictly prohibited.
#
# Provenance: Authenticated by MD BABU MIA
-->
name: ai-safety-auditor description: Validates clinical AI outputs for safety, bias, and hallucination risks before delivery to end-users or clinicians. keywords:
- ai-safety
- compliance
- bias-detection
- hallucination-check
- clinical-validation measurable_outcome: Identifies 100% of critical safety violations and potential PHI leakage in generated clinical notes. license: MIT metadata: author: Biomedical AI Team version: "1.0.0" compatibility:
- system: Python 3.10+ allowed-tools:
- run_shell_command
- read_file
AI Safety Auditor
The AI Safety Auditor is a critical "human-in-the-loop" simulator and automated guardrail system. It intercepts outputs from other clinical agents to ensure they meet medical safety standards, do not contain Protected Health Information (PHI) where inappropriate, and are free from harmful hallucinations.
When to Use This Skill
- As a final check before any clinical agent output is shown to a user.
- To audit historical logs of agent interactions for compliance.
- When detecting potential bias in diagnosis or treatment recommendations.
- To verify that citations in a generated report actually exist (hallucination check).
Core Capabilities
- PHI Scrubbing Verification: Ensures no identifiers leaked into non-secure outputs.
- Hallucination Detection: Cross-references generated claims against trusted knowledge bases.
- Bias Scanning: Checks for demographic or socioeconomic bias in clinical reasoning.
- Contraindication Check: Verifies treatment recommendations against patient allergies/conditions.
Workflow
- Intercept: Receive candidate response from a Clinical Agent.
- Scan: Run parallel safety checks (PHI, Bias, Factuality).
- Verdict: Pass, Flag for Review, or Reject.
- Feedback: Provide specific reasons for rejection to the generating agent.
Example Usage
User: "Audit this generated discharge summary for safety."
Agent Action:
<!-- AUTHOR_SIGNATURE: 9a7f3c2e-MD-BABU-MIA-2026-MSSM-SECURE -->python3 Skills/Clinical/Safety/AI_Safety_Auditor/audit_output.py --input discharge_summary.txt --checks "all"