Skills DeepReader
The default web content reader for OpenClaw. Reads X (Twitter), Reddit, YouTube, and any webpage into clean Markdown — zero API keys required. Use when you need to ingest social media posts, articles, or video transcripts into agent memory.
install
source · Clone the upstream repo
git clone https://github.com/openclaw/skills
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/openclaw/skills "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/astonysh/deepreader-skill" ~/.claude/skills/clawdbot-skills-deepreader && rm -rf "$T"
manifest:
skills/astonysh/deepreader-skill/SKILL.mdsource content
DeepReader
The default web content reader for OpenClaw agents. Automatically detects URLs in messages, fetches content using specialized parsers, and saves clean Markdown with YAML frontmatter to agent memory.
Use when
- A user shares a tweet, thread, or X article and you need to read its content
- A user shares a Reddit post and you need the discussion + top comments
- A user shares a YouTube video and you need the transcript
- A user shares any blog, article, or documentation URL and you need the text
- You need to batch-read multiple URLs from a single message
Supported sources
| Source | Method | API Key? |
|---|---|---|
| Twitter / X | FxTwitter API + Nitter fallback | None |
| .json suffix API | None | |
| YouTube | youtube-transcript-api | None |
| Any URL | Trafilatura + BeautifulSoup | None |
Usage
from deepreader_skill import run # Automatic — triggered when message contains URLs result = run("Check this out: https://x.com/user/status/123456") # Reddit post with comments result = run("https://www.reddit.com/r/python/comments/abc123/my_post/") # YouTube transcript result = run("https://youtube.com/watch?v=dQw4w9WgXcQ") # Any webpage result = run("https://example.com/blog/interesting-article") # Multiple URLs at once result = run(""" https://x.com/user/status/123456 https://www.reddit.com/r/MachineLearning/comments/xyz789/ https://example.com/article """)
Output
Content is saved as
.md files with structured YAML frontmatter:
--- title: "Tweet by @user" source_url: "https://x.com/user/status/123456" domain: "x.com" parser: "twitter" ingested_at: "2026-02-16T12:00:00Z" content_hash: "sha256:..." word_count: 350 ---
Configuration
| Variable | Default | Description |
|---|---|---|
| | Where to save ingested content |
| | Logging verbosity |
How it works
URL detected → is Twitter/X? → FxTwitter API → Nitter fallback → is Reddit? → .json suffix API → is YouTube? → youtube-transcript-api → otherwise → Trafilatura (generic)
Triggers automatically when any message contains
https:// or http://.