Claude-skill-registry content-filter
Filter and classify AI research content for relevance. Use when processing raw content from Twitter, Substacks, blogs, or podcasts to determine if it's worth extracting claims from. Assigns relevance scores, topics, and author categories.
install
source · Clone the upstream repo
git clone https://github.com/majiayu000/claude-skill-registry
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/majiayu000/claude-skill-registry "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/data/HypeDelta" ~/.claude/skills/majiayu000-claude-skill-registry-content-filter && rm -rf "$T"
manifest:
skills/data/HypeDelta/SKILL.mdsource content
Content Filter Skill
Assess content for relevance to AI research intelligence gathering. Filter noise and classify what remains.
Assessment Criteria
1. Relevance Score (0.0-1.0)
How relevant is this to understanding AI research progress, capabilities, limitations, or field direction?
| Score Range | Meaning | Examples |
|---|---|---|
| 0.0-0.3 | Not relevant | Personal updates, off-topic, promotional |
| 0.3-0.6 | Tangentially relevant | General tech news, adjacent topics |
| 0.6-0.8 | Relevant | Discusses AI research, capabilities, field |
| 0.8-1.0 | Highly relevant | Substantive claims, predictions, research insights |
2. Topic Classification
Assign ONE primary topic:
: Scaling laws, compute, training efficiencyscaling
: LLM reasoning, chain-of-thought, planning capabilitiesreasoning
: AI agents, tool use, autonomyagents
: AI safety, alignment, controlsafety
: Mechanistic interpretability, understanding modelsinterpretability
: Vision, audio, video modelsmultimodal
: RLHF, preference learning, Constitutional AIrlhf
: Embodied AI, roboticsrobotics
: Evals, benchmarks, capability measurementbenchmarks
: Training infra, chips, hardwareinfrastructure
: AI policy, regulation, governancepolicy
: General AI commentarygeneral
: Doesn't fit above categoriesother
3. Content Type
What kind of content is this?
: Makes claims about future AI capabilities/timelinesprediction
: Hints at ongoing/unpublished researchresearch-hint
: Expresses opinion on AI progress/directionopinion
: Reports factual information about released workfactual
: Critiques AI capabilities or claimscritique
: Meta-commentary on the fieldmeta
: Not substantivenoise
4. Substantiveness
Does this contain actual claims, arguments, or insights?
Substantive examples:
- "We found that CoT prompting shows diminishing returns beyond 8 steps"
- "The next generation will likely solve ARC-AGI"
- "Interpretability research is underrated"
Non-substantive examples:
- "Cool paper!" (reaction only)
- "Link: [url]" (link share without commentary)
- "Having coffee ☕" (personal update)
5. Author Category
Classify the author:
: Works at major AI lab (Anthropic, OpenAI, DeepMind, Meta AI, xAI, Mistral, Cohere)lab-researcher
: Known AI skeptic/critic with credentials (Marcus, Chollet, Mitchell, Bender, Brooks)critic
: University researcheracademic
: Independent researcher/commentatorindependent
: AI journalistjournalist
: Cannot determineunknown
Output Format
Return JSON:
{ "assessments": [ { "itemIndex": 0, "relevance": 0.85, "topic": "reasoning", "contentType": "research-hint", "isSubstantive": true, "authorCategory": "lab-researcher", "brief": "One sentence summary" } ] }
Filtering Heuristics
High Signal Indicators
- Lab researchers discussing their own work area
- Specific technical claims with numbers/benchmarks
- Predictions with timeframes
- Explicit disagreements between notable figures
- Hints using hedged language ("we've been seeing...", "I can't say much but...")
Low Signal Indicators
- Pure link shares without commentary
- Conference attendance announcements
- Hiring posts
- Generic congratulations
- Retweets without quote
- Personal life updates
- Product launches (unless with technical claims)
Gray Areas
- Paper summaries (relevant if includes opinion/analysis)
- Q&A responses (depends on question depth)
- Thread continuations (may need full thread context)