Claude-skill-registry comp-scout-scrape
Scrape competition websites, extract structured data, and auto-persist to GitHub issues. Creates issues for new competitions, adds comments for duplicates.
git clone https://github.com/majiayu000/claude-skill-registry
T=$(mktemp -d) && git clone --depth=1 https://github.com/majiayu000/claude-skill-registry "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/data/comp-scout-scrape" ~/.claude/skills/majiayu000-claude-skill-registry-comp-scout-scrape && rm -rf "$T"
skills/data/comp-scout-scrape/SKILL.mdCompetition Scraper
Scrape creative writing competitions from Australian aggregator sites and automatically persist to GitHub.
What This Skill Does
- Scrapes competitions.com.au and netrewards.com.au
- Extracts structured data (dates, prompts, prizes)
- Checks for duplicates against existing GitHub issues (by URL and title similarity)
- Creates issues for NEW competitions only
- Adds comments to existing issues when same competition found on another site
- Skips competitions that are already tracked
The scraper already filters out sponsored/lottery ads. Your job is to check for duplicates, then persist only new competitions.
What Counts as "New"
A competition is NEW if:
- Its URL is not found in any existing issue body (check the full body text, not just the primary URL field)
- AND its normalized title is <80% similar to all existing issue titles
A competition is a DUPLICATE if:
- Its URL appears anywhere in an existing issue (body text, comments) → already tracked, skip
- Its normalized title is >80% similar to an existing issue title → likely same competition, skip
- Same competition found on a different aggregator site → add comment to existing issue noting the alternate URL
Note: An issue body may contain multiple URLs (one per aggregator site). When checking for duplicates, search the entire issue body for the scraped URL, not just a specific field.
Word Limit Clarification
"25WOL" is a category name, NOT a filter. Competitions with 25, 50, or 100 word limits are all valid creative writing competitions - persist them all (if new).
Prerequisites
pip install playwright playwright install chromium
Also requires:
CLI authenticatedgh- Target repository for competition data (not this skills repo)
Workflow
Step 1: Determine Target Repository
The target repo stores competition issues. Specify or get from config:
# From workspace config (if hiivmind-pulse-gh initialized) TARGET_REPO=$(yq '.repositories[0].full_name' .hiivmind/github/config.yaml 2>/dev/null) # Or use default/specified TARGET_REPO="${TARGET_REPO:-discreteds/competition-data}"
Step 2: Scrape Listings
Run the scraper to get structured competition data:
python skills/comp-scout-scrape/scraper.py listings
Output:
{ "competitions": [ { "url": "https://competitions.com.au/win-example/", "site": "competitions.com.au", "title": "Win a $500 Gift Card", "normalized_title": "500 gift card", "brand": "Example Brand", "prize_summary": "$500", "prize_value": 500, "closing_date": "2024-12-31" } ], "scrape_date": "2024-12-09", "errors": [] }
Step 3: Check for Existing Issues
For each scraped competition, check if it already exists:
# Get all open competition issues gh issue list -R "$TARGET_REPO" \ --label "competition" \ --state open \ --json number,title,body \ --limit 200
Match by:
- URL in issue body (exact match = definite duplicate)
- Normalized title similarity (>80% = likely duplicate)
Step 4: Fetch Details for New Competitions
For competitions not already tracked, get full details:
python skills/comp-scout-scrape/scraper.py detail "https://competitions.com.au/win-example/"
For multiple new competitions, use batch mode:
echo '{"urls": ["url1", "url2", ...]}' | python skills/comp-scout-scrape/scraper.py details-batch
Step 4.5: Apply Auto-Tagging Rules (NOT Filtering)
IMPORTANT: Auto-tagging is for LABELING issues, not for skipping/excluding competitions.
Check competitions against user preferences from the data repo's CLAUDE.md to determine which labels to apply.
- Fetch preferences:
gh api repos/$TARGET_REPO/contents/CLAUDE.md -H "Accept: application/vnd.github.raw" 2>/dev/null
-
Parse the Detection Keywords section for tagging rules
-
For each competition, check if title/prize matches any keywords:
For each tag_rule in [for-kids, cruise]: For each keyword in tag_rule.keywords: If keyword.lower() in (competition.title + competition.prize_summary).lower(): Add tag_rule.label to issue labels
- ALL competitions are ALWAYS persisted as issues. Tagged competitions:
- Get the relevant label applied (e.g.,
,for-kids
)cruise - Are closed immediately with explanation comment
- But they ARE STILL CREATED as issues (for record-keeping and potential review)
- Get the relevant label applied (e.g.,
Step 5: Auto-Persist Results
For New Competitions → Create Issue
gh issue create -R "$TARGET_REPO" \ --title "$TITLE" \ --label "competition" \ --label "25wol" \ --body "$(cat <<'EOF' ## Competition Details **URL:** {url} **Brand:** {brand} **Prize:** {prize_summary} **Word Limit:** {word_limit} words **Closes:** {closing_date} **Draw Date:** {draw_date} **Winners Notified:** {notification_info} ## Prompt > {prompt} --- *Scraped from {site} on {scrape_date}* EOF )"
Then set milestone by closing month:
gh issue edit $ISSUE_NUMBER -R "$TARGET_REPO" --milestone "December 2024"
For Duplicates → Add Comment
If competition URL found on another site:
gh issue comment $EXISTING_ISSUE -R "$TARGET_REPO" --body "$(cat <<'EOF' ### Also found on {other_site} **URL:** {url} **Title on this site:** {title} *Discovered: {date}* EOF )"
For Filtered Competitions → Create Issue + Close
If competition matched auto-filter keywords:
# Create the issue first (for record-keeping) ISSUE_URL=$(gh issue create -R "$TARGET_REPO" \ --title "$TITLE" \ --label "competition" \ --label "25wol" \ --label "$FILTER_LABEL" \ --body "...") # Extract issue number ISSUE_NUMBER=$(echo "$ISSUE_URL" | grep -oE '[0-9]+$') # Close with explanation gh issue close $ISSUE_NUMBER -R "$TARGET_REPO" --comment "$(cat <<'EOF' Auto-filtered: matches '$KEYWORD' in $FILTER_RULE preferences. See CLAUDE.md in this repository for filter settings. EOF )"
Step 6: Report Results
Present confirmation to user:
✅ Scrape complete! **Created 3 new issues:** - #42: Win a $500 Coles Gift Card (closes Dec 31) - #43: Win a Trip to Bali (closes Jan 15) - #44: Win a Year's Supply of Coffee (closes Dec 20) **Auto-filtered 2 (created + closed):** - #45: Win Lego Set (for-kids: matched "Lego") - #46: Win P&O Cruise (cruise: matched "P&O") **Found 2 duplicates (added as comments):** - #38: Win Woolworths Gift Cards (also on netrewards.com.au) - #39: Win Dreamworld Experience (also on netrewards.com.au) **Skipped 7 already tracked**
IMPORTANT: Do NOT ask "Would you like me to analyze these?" at the end. When invoked by
comp-scout-daily, the workflow will automatically invoke analyze/compose skills next. Report results and stop.
Output Fields
Listing Output
| Field | Type | Description |
|---|---|---|
| url | string | Full URL to competition detail page |
| site | string | Source site (competitions.com.au or netrewards.com.au) |
| title | string | Competition title as displayed |
| normalized_title | string | Lowercase, prefixes stripped, for matching |
| brand | string | Sponsor/brand name (if available) |
| prize_summary | string | Prize description or value badge |
| prize_value | int/null | Numeric value in dollars |
| closing_date | string/null | YYYY-MM-DD format |
Detail Output
All listing fields plus:
| Field | Type | Description |
|---|---|---|
| prompt | string | The actual competition question/prompt |
| word_limit | int | Maximum words (default 25) |
| entry_method | string | How to submit entry |
| winner_notification | object/null | Notification details from JSON-LD |
| scraped_at | string | ISO timestamp of scrape |
Winner Notification Object
| Field | Type | Description |
|---|---|---|
| notification_text | string | Raw notification text |
| notification_date | string/null | Specific date if mentioned |
| notification_days | int/null | Days after close/draw |
| selection_text | string | How winners are selected |
| selection_date | string/null | When judging occurs |
Title Normalization
Titles are normalized for deduplication:
- Lowercase
- Strip prefixes: "Win ", "Win a ", "Win an ", "Win the ", "Win 1 of "
- Remove punctuation
- Collapse whitespace
Example:
Original: "Win a $500 Coles Gift Card" Normalized: "500 coles gift card"
Example Session
User: Scrape competitions Claude: I'll scrape competitions and persist new ones to GitHub. [Runs: python skills/comp-scout-scrape/scraper.py listings] Found 12 competitions from both sites. [Runs: gh issue list -R discreteds/competition-data --label competition --json number,title,body] Checking against 45 existing issues... - 3 are new - 2 are duplicates (same competition, different source) - 7 already tracked Fetching details for 3 new competitions... [Creates issues and adds comments] ✅ Scrape complete! **Created 3 new issues:** - #46: Win a $500 Coles Gift Card (closes Dec 31) - Milestone: December 2024 - #47: Win a Trip to Bali (closes Jan 15) - Milestone: January 2025 - #48: Win a Year's Supply of Coffee (closes Dec 20) - Milestone: December 2024 **Added 2 duplicate comments:** - #38: Also found on netrewards.com.au - #39: Also found on netrewards.com.au
CLI Commands Reference
# Scrape all listing pages python skills/comp-scout-scrape/scraper.py listings # Get full details for one competition python skills/comp-scout-scrape/scraper.py detail "URL" # Get full details for multiple competitions (batch mode) echo '{"urls": ["url1", "url2"]}' | python skills/comp-scout-scrape/scraper.py details-batch # Debug: just get URLs python skills/comp-scout-scrape/scraper.py urls
Batch Details Output
{ "details": [ { "url": "...", "title": "...", "prompt": "Tell us in 25 words...", "word_limit": 25, ... } ], "scrape_date": "2024-12-09", "errors": [] }
Persistence Details
This skill handles all GitHub persistence. The separate
comp-scout-persist skill is deprecated - its functionality is merged here.
Issue Creation Template
## Competition Details **URL:** {url} **Brand:** {brand} **Prize:** {prize_summary} **Word Limit:** {word_limit} words **Closes:** {closing_date} **Draw Date:** {draw_date} **Winners Notified:** {notification_info} ## Prompt > {prompt} --- *Scraped from {site} on {scrape_date}*
Labels
| Label | Description | Auto-applied |
|---|---|---|
| All competition issues | Always |
| 25 words or less type | Always |
| Auto-filtered (kids competitions) | When keyword matches |
| Auto-filtered (cruise competitions) | When keyword matches |
| Closes within 3 days | By separate check |
| Entry has been composed | By comp-scout-compose |
| Entry has been submitted | Manually |
Milestones
Issues are assigned to milestones by closing date month:
- "December 2024"
- "January 2025"
- etc.
# Create milestone if needed gh api repos/$TARGET_REPO/milestones \ --method POST \ --field title="$MONTH_YEAR" \ --field due_on="$LAST_DAY_OF_MONTH" # Assign to issue gh issue edit $ISSUE_NUMBER -R "$TARGET_REPO" --milestone "$MONTH_YEAR"
Duplicate Comment Template
### Also found on {other_site} **URL:** {url} **Title on this site:** {title} *Discovered: {date}*
Filtered Issue Handling
When a competition matches filter keywords:
- Issue is created (for record-keeping)
- Filter label is applied (e.g.,
)for-kids - Issue is immediately closed with explanation
gh issue close $ISSUE_NUMBER -R "$TARGET_REPO" \ --comment "Auto-filtered: matches '$KEYWORD' in $FILTER_RULE preferences."
Integration
This skill is invoked by
comp-scout-daily as the first step in the workflow.
After scraping, you can:
- Use comp-scout-analyze to generate entry strategies
- Use comp-scout-compose to write actual entries
- Both will auto-persist their results as comments on the issue