Skills local-whisper

Local speech-to-text using OpenAI Whisper. Runs fully offline after model download. High quality transcription with multiple model sizes.

install
source · Clone the upstream repo
git clone https://github.com/openclaw/skills
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/openclaw/skills "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/araa47/local-whisper" ~/.claude/skills/clawdbot-skills-local-whisper && rm -rf "$T"
manifest: skills/araa47/local-whisper/SKILL.md
source content

Local Whisper STT

Local speech-to-text using OpenAI's Whisper. Fully offline after initial model download.

Usage

# Basic
~/.clawdbot/skills/local-whisper/scripts/local-whisper audio.wav

# Better model
~/.clawdbot/skills/local-whisper/scripts/local-whisper audio.wav --model turbo

# With timestamps
~/.clawdbot/skills/local-whisper/scripts/local-whisper audio.wav --timestamps --json

Models

ModelSizeNotes
tiny
39MFastest
base
74MDefault
small
244MGood balance
turbo
809MBest speed/quality
large-v3
1.5GBMaximum accuracy

Options

  • --model/-m
    — Model size (default: base)
  • --language/-l
    — Language code (auto-detect if omitted)
  • --timestamps/-t
    — Include word timestamps
  • --json/-j
    — JSON output
  • --quiet/-q
    — Suppress progress

Setup

Uses uv-managed venv at

.venv/
. To reinstall:

cd ~/.clawdbot/skills/local-whisper
uv venv .venv --python 3.12
uv pip install --python .venv/bin/python click openai-whisper torch --index-url https://download.pytorch.org/whl/cpu