Awesome-omni-skill kimmo-agent-friendly-score

Score developer tools and SaaS products for AI agent compatibility. Use when evaluating how well a devtool works with AI coding assistants, or when optimizing a product for the agent era.

install
source · Clone the upstream repo
git clone https://github.com/diegosouzapw/awesome-omni-skill
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/diegosouzapw/awesome-omni-skill "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/data-ai/kimmo-agent-friendly-score-ihmissuti" ~/.claude/skills/diegosouzapw-awesome-omni-skill-kimmo-agent-friendly-score && rm -rf "$T"
manifest: skills/data-ai/kimmo-agent-friendly-score-ihmissuti/SKILL.md
source content

Agent-Friendly Score - DevTool Evaluation

Evaluate developer tools and SaaS products for compatibility with AI coding assistants (Cursor, Claude, GitHub Copilot).

When to Use

  • User asks "How agent-friendly is [tool]?"
  • User wants to evaluate a devtool for AI compatibility
  • User is building a devtool and wants to optimize for AI assistants
  • User is comparing tools and AI compatibility matters

The Agent Era Context

85% of developers now use AI tools regularly. When a developer asks Cursor to "add email functionality," the AI picks the service, writes the integration, and runs the install command.

The New Funnel:

  • Traditional: Marketing → Landing → Docs → Trial → Conversion
  • Agent: Problem → AI suggestion →
    npm install
    → Subscription

Tools that AI can easily work with get recommended. Tools it can't work with become invisible.

Scoring Framework

Category 1: SDK & API Design (30 points)

CriterionPointsHow to Check
SDK available for major languages10Check docs for JS, Python, Go, etc.
Consistent, predictable API10Review API reference for patterns
Complete TypeScript definitions5Check npm package for .d.ts files
Clear error messages5Test error responses

Scoring guide:

  • 25-30: Excellent - AI can generate correct code first try
  • 15-24: Good - AI mostly succeeds, occasional fixes needed
  • 0-14: Poor - AI struggles to generate working code

Category 2: Documentation Quality (25 points)

CriterionPointsHow to Check
Docs lead with working code10First thing on quickstart page
Copy-paste examples work5Try the first 3 examples
Parseable structure (H1→H2→H3)5View page source/outline
No login walls on docs5Access docs without account

Scoring guide:

  • 20-25: AI can extract and apply correctly
  • 10-19: AI needs some interpretation
  • 0-9: AI will likely hallucinate or fail

Category 3: Training Data Presence (20 points)

CriterionPointsHow to Check
GitHub repos using this tool8Search GitHub for imports
Stack Overflow presence6Search SO for [tool] questions
Tutorial/blog coverage6Search "[tool] tutorial"

Scoring guide:

  • 15-20: Strong training data signal
  • 8-14: Moderate presence
  • 0-7: AI may not know this tool well

Category 4: MCP Integration (15 points)

CriterionPointsHow to Check
Official MCP server exists10Check mcp.so, official docs
MCP server is maintained3Recent commits, version updates
MCP server is discoverable2Listed on MCP.so or npm

Scoring guide:

  • 12-15: Full agent workflow integration
  • 5-11: Partial integration
  • 0-4: Not in agent workflow

Category 5: Time to Working (10 points)

CriterionPointsHow to Check
Install to "hello world" <5 min5Time yourself following quickstart
No complex onboarding3Can start without account?
Sensible defaults2Works without config?

Scoring guide:

  • 8-10: Instant productivity
  • 4-7: Reasonable setup
  • 0-3: Significant friction

Evaluation Workflow

Step 1: Identify the Tool

Get from user:

  • Tool name and URL
  • Category (email, auth, database, etc.)
  • Main competitor to compare against

Step 2: SDK Evaluation

Check official SDK:

# Check npm for TypeScript types
npm info [package] types

# Check for SDK in multiple languages
# Visit: github.com/[org] and look for SDK repos

Test API consistency:

  • Are endpoints predictable? (e.g.,
    /users
    ,
    /users/:id
    )
  • Are responses consistent?
  • Are errors structured?

Step 3: Documentation Audit

Visit docs and check:

  • First code example is within scroll view
  • Examples include all necessary imports
  • Examples actually work when copied
  • Structure uses semantic headings
  • No authentication required to view

Step 4: Training Data Check

Search GitHub:

"import { X } from '[package]'" language:JavaScript
"from [package] import" language:Python

Search Stack Overflow:

[tool] is:question

Step 5: MCP Check

Search for MCP server:

  • https://mcp.so - search for tool name
  • Official docs - search for "MCP" or "Model Context Protocol"
  • GitHub - search "[tool] mcp server"

Step 6: Time Test

Follow quickstart:

  1. Start timer
  2. Follow official quickstart exactly
  3. Stop when first API call succeeds
  4. Record time and friction points

Output Template

# Agent-Friendly Score: [Tool Name]

## Overall Score: [X]/100

| Category         | Score | Max |
| ---------------- | ----- | --- |
| SDK & API Design | X     | 30  |
| Documentation    | X     | 25  |
| Training Data    | X     | 20  |
| MCP Integration  | X     | 15  |
| Time to Working  | X     | 10  |

## Grade: [A/B/C/D/F]

- A (85-100): AI will recommend and integrate correctly
- B (70-84): AI will usually succeed
- C (55-69): AI needs help, may hallucinate
- D (40-54): Significant AI compatibility issues
- F (<40): AI will struggle or avoid

## Breakdown

### SDK & API Design ([X]/30)

**Strengths:**

- [What works well]

**Gaps:**

- [What's missing]

### Documentation ([X]/25)

**Strengths:**

- [What works well]

**Gaps:**

- [What's missing]

### Training Data Presence ([X]/20)

- GitHub repos found: [X]
- Stack Overflow questions: [X]
- Tutorial coverage: [High/Medium/Low]

### MCP Integration ([X]/15)

- MCP server: [Official/Community/None]
- Status: [Active/Stale/N/A]

### Time to Working ([X]/10)

- Quickstart time: [X minutes]
- Friction points: [list]

## Recommendations

### Quick Wins (High Impact, Low Effort)

1. [Recommendation]
2. [Recommendation]

### Strategic Improvements

1. [Recommendation]
2. [Recommendation]

## Competitor Comparison

| Metric          | [Tool] | [Competitor] |
| --------------- | ------ | ------------ |
| Agent Score     | X/100  | Y/100        |
| MCP Server      | Yes/No | Yes/No       |
| Time to Working | X min  | Y min        |

## Verdict

[One paragraph summary of whether this tool is positioned for the agent era]

Benchmarks by Category

Email APIs

ToolTypical ScoreNotes
Resend85-90MCP, clean SDK, great docs
Postmark80-85MCP, enterprise-ready
SendGrid60-70No official MCP, legacy API

Authentication

ToolTypical ScoreNotes
Clerk85-90MCP, great DX
Auth075-80MCP, but complex
Firebase Auth80-85MCP, Google ecosystem

Databases

ToolTypical ScoreNotes
Supabase85-90MCP, hosted, great DX
Neon85-90MCP, serverless
PlanetScale75-80MCP (read-only)

Use these as calibration when scoring.

Key Insight

The best technical product doesn't always win anymore. The most AI-accessible product wins. When an AI assistant can:

  1. Understand your docs
  2. Generate working code
  3. Integrate via MCP

...you're in the conversation. When it can't, you're invisible to the fastest-growing developer segment.


By Kimmo Ihanus | kimmoihanus.com