Skills local-whisper
Local speech-to-text using OpenAI Whisper. Runs fully offline after model download. High quality transcription with multiple model sizes.
install
source · Clone the upstream repo
git clone https://github.com/openclaw/skills
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/openclaw/skills "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/araa47/local-whisper" ~/.claude/skills/clawdbot-skills-local-whisper && rm -rf "$T"
manifest:
skills/araa47/local-whisper/SKILL.mdsource content
Local Whisper STT
Local speech-to-text using OpenAI's Whisper. Fully offline after initial model download.
Usage
# Basic ~/.clawdbot/skills/local-whisper/scripts/local-whisper audio.wav # Better model ~/.clawdbot/skills/local-whisper/scripts/local-whisper audio.wav --model turbo # With timestamps ~/.clawdbot/skills/local-whisper/scripts/local-whisper audio.wav --timestamps --json
Models
| Model | Size | Notes |
|---|---|---|
| 39M | Fastest |
| 74M | Default |
| 244M | Good balance |
| 809M | Best speed/quality |
| 1.5GB | Maximum accuracy |
Options
— Model size (default: base)--model/-m
— Language code (auto-detect if omitted)--language/-l
— Include word timestamps--timestamps/-t
— JSON output--json/-j
— Suppress progress--quiet/-q
Setup
Uses uv-managed venv at
.venv/. To reinstall:
cd ~/.clawdbot/skills/local-whisper uv venv .venv --python 3.12 uv pip install --python .venv/bin/python click openai-whisper torch --index-url https://download.pytorch.org/whl/cpu