Claude-skill-registry extract-hyperparameters
Identify and document model hyperparameters from papers. Use when setting up training configurations.
install
source · Clone the upstream repo
git clone https://github.com/majiayu000/claude-skill-registry
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/majiayu000/claude-skill-registry "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/data/extract-hyperparameters" ~/.claude/skills/majiayu000-claude-skill-registry-extract-hyperparameters && rm -rf "$T"
manifest:
skills/data/extract-hyperparameters/SKILL.mdsource content
Extract Hyperparameters
Locate and document all hyperparameters mentioned in research papers including learning rates, batch sizes, and model configurations.
When to Use
- Reproducing paper results
- Setting up model training configurations
- Comparing hyperparameter choices across papers
- Planning hyperparameter tuning experiments
Quick Reference
# Extract numeric values and parameters from papers pdftotext paper.pdf - | grep -i "learning rate\|batch\|epochs\|weight decay\|dropout" | head -20 # Common pattern search grep -E "\\b(lr|batch_size|epochs|momentum|dropout|layers)\\s*[=:]" config.py
Workflow
- Find hyperparameter table: Look for "Table 1" or "Hyperparameters" section
- Document architecture parameters: Layer sizes, activation functions, normalization
- Extract training parameters: Learning rate, batch size, epochs, optimizers
- Note regularization: Dropout, weight decay, batch normalization
- Create configuration file: Translate to implementation format (YAML/JSON/Mojo)
Output Format
Hyperparameter documentation:
- Model architecture (layers, sizes, activations)
- Training parameters (LR, batch size, epochs)
- Optimizer configuration (type, momentum, decay)
- Regularization settings (dropout, L1/L2)
- Data preprocessing (normalization, augmentation)
- Hardware and precision (float32, float64)
References
- See
skill for data configurationprepare-dataset - See
skill for training implementationtrain-model - See
for Mojo configuration patterns/notes/review/mojo-ml-patterns.md