Skills ocr-local

Extract text from images using Tesseract.js OCR (100% local, no API key required). Supports Chinese (simplified/traditional) and English.

install
source · Clone the upstream repo
git clone https://github.com/openclaw/skills
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/openclaw/skills "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/15914355527/ocr-local-v2" ~/.claude/skills/openclaw-skills-ocr-local && rm -rf "$T"
OpenClaw · Install into ~/.openclaw/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/openclaw/skills "$T" && mkdir -p ~/.openclaw/skills && cp -r "$T/skills/15914355527/ocr-local-v2" ~/.openclaw/skills/openclaw-skills-ocr-local && rm -rf "$T"
manifest: skills/15914355527/ocr-local-v2/SKILL.md
source content

OCR - Image Text Recognition (Local)

Extract text from images using Tesseract.js. 100% local run, no API key required. Supports Chinese and English.

Quick start

node {baseDir}/scripts/ocr.js /path/to/image.jpg
node {baseDir}/scripts/ocr.js /path/to/image.png --lang chi_sim
node {baseDir}/scripts/ocr.js /path/to/image.jpg --lang chi_tra+eng

Options

  • --lang <langs>
    : Language codes (default: chi_sim+eng)

    • chi_sim
      - Simplified Chinese
    • chi_tra
      - Traditional Chinese
    • eng
      - English
    • Combine with
      +
      :
      chi_sim+eng
  • --json
    : Output as JSON instead of plain text

Examples

# Recognize Chinese screenshot
node {baseDir}/scripts/ocr.js screenshot.png

# Recognize English document
node {baseDir}/scripts/ocr.js document.jpg --lang eng

# Mixed Chinese + English
node {baseDir}/scripts/ocr.js mixed.png --lang chi_sim+eng

Notes

  • First run downloads language data (~20MB per language)
  • Subsequent runs are cached locally
  • Works best with clear, high-contrast images
  • For handwritten text, accuracy may vary