Skills grant-budget-justification

Generate narrative budget justifications for NIH/NSF applications

install
source · Clone the upstream repo
git clone https://github.com/openclaw/skills
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/openclaw/skills "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/aipoch-ai/grant-budget-justification" ~/.claude/skills/openclaw-skills-grant-budget-justification && rm -rf "$T"
OpenClaw · Install into ~/.openclaw/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/openclaw/skills "$T" && mkdir -p ~/.openclaw/skills && cp -r "$T/skills/aipoch-ai/grant-budget-justification" ~/.openclaw/skills/openclaw-skills-grant-budget-justification && rm -rf "$T"
manifest: skills/aipoch-ai/grant-budget-justification/SKILL.md
source content

Grant Budget Justification

Narrative budget explanations for grant proposals.

Use Cases

  • Equipment purchases
  • Personnel costs
  • Supplies and reagents
  • Travel and dissemination

Parameters

ParameterTypeDefaultRequiredDescription
--input
,
-i
string-YesPath to budget items file (JSON/CSV)
--justification-type
string-YesType of justification (Equipment, Personnel, Other)
--agency
stringNIHNoFunding agency (NIH, NSF)
--output
,
-o
stringstdoutNoOutput file path
--format
stringtextNoOutput format (text, markdown, docx)

Returns

  • Narrative justification text
  • Cost-benefit rationale
  • Compliance with agency requirements

Example

Input: $50,000 for mass spectrometer Output: Justification emphasizing essentiality and cost-sharing

Risk Assessment

Risk IndicatorAssessmentLevel
Code ExecutionPython/R scripts executed locallyMedium
Network AccessNo external API callsLow
File System AccessRead input files, write output filesMedium
Instruction TamperingStandard prompt guidelinesLow
Data ExposureOutput files saved to workspaceLow

Security Checklist

  • No hardcoded credentials or API keys
  • No unauthorized file system access (../)
  • Output does not expose sensitive information
  • Prompt injection protections in place
  • Input file paths validated (no ../ traversal)
  • Output directory restricted to workspace
  • Script execution in sandboxed environment
  • Error messages sanitized (no stack traces exposed)
  • Dependencies audited

Prerequisites

No additional Python packages required.

Evaluation Criteria

Success Metrics

  • Successfully executes main functionality
  • Output meets quality standards
  • Handles edge cases gracefully
  • Performance is acceptable

Test Cases

  1. Basic Functionality: Standard input → Expected output
  2. Edge Case: Invalid input → Graceful error handling
  3. Performance: Large dataset → Acceptable processing time

Lifecycle Status

  • Current Stage: Draft
  • Next Review Date: 2026-03-06
  • Known Issues: None
  • Planned Improvements:
    • Performance optimization
    • Additional feature support