decodo-scraper
Search Google, scrape web pages, Amazon product pages, YouTube subtitles, or Reddit (post/subreddit) using the Decodo Scraper OpenClaw Skill.
git clone https://github.com/Decodo/decodo-openclaw-skill
git clone --depth=1 https://github.com/Decodo/decodo-openclaw-skill ~/.claude/skills/decodo-decodo-openclaw-skill-decodo-scraper
git clone --depth=1 https://github.com/Decodo/decodo-openclaw-skill ~/.openclaw/skills/decodo-decodo-openclaw-skill-decodo-scraper
SKILL.mdDecodo Scraper OpenClaw Skill
Use this skill to search Google, scrape any URL, or fetch YouTube subtitles via the Decodo Web Scraping API. Search outputs a JSON object of result sections; Scrape URL outputs plain markdown; Amazon and Amazon search output parsed product-page or search results (JSON). Amazon search uses
--query. YouTube subtitles outputs transcript/subtitles. Reddit post and Reddit subreddit output post/listing content (JSON).
Authentication: Set
DECODO_AUTH_TOKEN (Basic auth token from Decodo Dashboard → Scraping APIs) in your environment or in a .env file in the repo root.
Errors: On failure the script writes a JSON error to stderr and exits with code 1.
Tools
1. Search Google
Use this to find URLs, answers, or structured search results. The API returns a JSON object whose
results key contains several sections (not all may be present for every query):
| Section | Description |
|---|---|
| Main search results (titles, links, snippets). |
| AI-generated overviews or summaries when Google shows them. |
| Paid/sponsored results (ads). |
| “People also ask”–style questions and answers. |
| Suggested related search queries. |
| Forum or discussion results (e.g. Reddit, Stack Exchange). |
The script outputs only the inner
results object (these sections); pagination info (page, last_visible_page, parse_status_code) is not included.
Command:
python3 tools/scrape.py --target google_search --query "your search query"
Examples:
python3 tools/scrape.py --target google_search --query "best laptops 2025" python3 tools/scrape.py --target google_search --query "python requests tutorial"
Optional:
--geo us or --locale en for location/language.
2. Scrape URL
Use this to get the content of a specific web page. By default the API returns content as Markdown (cleaner for LLMs and lower token usage).
Command:
python3 tools/scrape.py --target universal --url "https://example.com"
Examples:
python3 tools/scrape.py --target universal --url "https://example.com" python3 tools/scrape.py --target universal --url "https://news.ycombinator.com/"
3. Amazon product page
Use this to get parsed data from an Amazon product (or other Amazon) page. Pass the product page URL as
--url. The script sends parse: true and outputs the inner results object (e.g. ads, product details, etc.).
Command:
python3 tools/scrape.py --target amazon --url "https://www.amazon.com/dp/PRODUCT_ID"
Examples:
python3 tools/scrape.py --target amazon --url "https://www.amazon.com/dp/B09H74FXNW"
4. Amazon search
Use this to search Amazon and get parsed results (search results list, delivery_postcode, etc.). Pass the search query as
--query.
Command:
python3 tools/scrape.py --target amazon_search --query "your search query"
Examples:
python3 tools/scrape.py --target amazon_search --query "laptop"
5. YouTube subtitles
Use this to get subtitles/transcript for a YouTube video. Pass the video ID (e.g. from
youtube.com/watch?v=VIDEO_ID) as --query.
Command:
python3 tools/scrape.py --target youtube_subtitles --query "VIDEO_ID"
Examples:
python3 tools/scrape.py --target youtube_subtitles --query "dFu9aKJoqGg"
6. Reddit post
Use this to get the content of a Reddit post (thread). Pass the full post URL as
--url.
Command:
python3 tools/scrape.py --target reddit_post --url "https://www.reddit.com/r/SUBREDDIT/comments/ID/..."
Examples:
python3 tools/scrape.py --target reddit_post --url "https://www.reddit.com/r/nba/comments/17jrqc5/serious_next_day_thread_postgame_discussion/"
7. Reddit subreddit
Use this to get the listing (posts) of a Reddit subreddit. Pass the subreddit URL as
--url.
Command:
python3 tools/scrape.py --target reddit_subreddit --url "https://www.reddit.com/r/SUBREDDIT/"
Examples:
python3 tools/scrape.py --target reddit_subreddit --url "https://www.reddit.com/r/nba/"
Summary
| Action | Target | Argument | Example command |
|---|---|---|---|
| Search | | | |
| Scrape page | | | |
| Amazon product | | | |
| Amazon search | | | |
| YouTube subtitles | | | |
| Reddit post | | | |
| Reddit subreddit | | | |
Output: Search → JSON (sections). Scrape URL → markdown. Amazon / Amazon search → JSON (results e.g. ads, product info, delivery_postcode). YouTube → transcript. Reddit → JSON (content).