Awesome-omni-skill crawl-docs-skill
Run a Crawl4AI-based doc crawler and save internal pages as Markdown using page titles as filenames. Use when the user provides a docs URL and wants all internal subpages saved as .md files. Environment setup should only use uv.
install
source · Clone the upstream repo
git clone https://github.com/diegosouzapw/awesome-omni-skill
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/diegosouzapw/awesome-omni-skill "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/documentation/crawl-docs-skill" ~/.claude/skills/diegosouzapw-awesome-omni-skill-crawl-docs-skill && rm -rf "$T"
manifest:
skills/documentation/crawl-docs-skill/SKILL.mdsource content
Crawl Docs Skill
One-shot (Direct Run)
- Run directly:
bash "<path-to-skill>/scripts/crawl_docs.sh" <URL> --out "$PWD/docs" - Do NOT install dependencies or scaffold files first; run immediately. The script will bootstrap uv + venv only if needed.
- Do NOT run tests or create plans; just execute the script and let it fail if the system is missing prerequisites (e.g. curl).
- Default output dir (skill convention):
(absolute path to"$PWD/docs"
under the current working directory); always pass it via./docs
unless the user specifies otherwise.--out - Optional flags:
,--out <dir>--max-pages <int>
Correct Usage (from real run)
- First run: execute the one-shot command exactly once with no extra steps.
- If it fails: fix the missing system prerequisite (most commonly
) and re-run the same command.curl - Do not probe versions or add max-pages unless the user asks.
Notes
- Use page title (tab name) as the output filename; auto-deduplicate with numeric suffixes.
- Default behavior has no max pages unless
is provided.--max-pages