Claude-code-plugins-plus-skills grammarly-incident-runbook
install
source · Clone the upstream repo
git clone https://github.com/jeremylongshore/claude-code-plugins-plus-skills
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/jeremylongshore/claude-code-plugins-plus-skills "$T" && mkdir -p ~/.claude/skills && cp -r "$T/plugins/saas-packs/grammarly-pack/skills/grammarly-incident-runbook" ~/.claude/skills/jeremylongshore-claude-code-plugins-plus-skills-grammarly-incident-runbook && rm -rf "$T"
manifest:
plugins/saas-packs/grammarly-pack/skills/grammarly-incident-runbook/SKILL.mdsource content
Grammarly Incident Runbook
Overview
Incident response procedures for Grammarly writing API integration failures. Covers text check timeouts, suggestion quality degradation, OAuth token failures, and rate limit storms. Grammarly powers real-time writing assistance, so API incidents directly impact user-facing text checking, scoring workflows, and content quality pipelines. Classify severity immediately using the matrix below and follow the corresponding playbook.
Severity Levels
| Level | Definition | Response Time | Example |
|---|---|---|---|
| P1 - Critical | Full API outage, all scoring requests fail | 15 min | 5xx on for all requests |
| P2 - High | OAuth token failures or sustained timeouts | 30 min | All authenticated requests return 401 |
| P3 - Medium | Rate limit storms or elevated latency | 2 hours | 429 responses, scoring takes 10s+ per request |
| P4 - Low | Suggestion quality drift or single endpoint issue | 8 hours | Scores returning but correctness values seem off |
Diagnostic Steps
# Test API health (unauthenticated) curl -s -o /dev/null -w "HTTP %{http_code}\n" \ https://api.grammarly.com/ecosystem/api/v2/scores # Test authenticated scoring curl -s -w "\nHTTP %{http_code}\n" \ -H "Authorization: Bearer $GRAMMARLY_ACCESS_TOKEN" \ -H "Content-Type: application/json" \ -X POST https://api.grammarly.com/ecosystem/api/v2/scores \ -d '{"text": "Test sentence for Grammarly API diagnostic health check."}' # Check OAuth token validity curl -s -o /dev/null -w "HTTP %{http_code}\n" \ -H "Authorization: Bearer $GRAMMARLY_ACCESS_TOKEN" \ https://api.grammarly.com/ecosystem/api/v2/account
Incident Playbooks
API Outage
- Confirm with unauthenticated health check (see diagnostics)
- Check Grammarly status page and developer announcements
- Activate fallback mode — return placeholder scores to avoid blocking users
- Queue text submissions for retry when API recovers
- Notify downstream consumers that scores are unavailable
Authentication Failure
- Test token validity with the account endpoint diagnostic above
- If 401: OAuth access token has expired — trigger token refresh flow
- If refresh token also fails: re-authorize via OAuth consent flow
- Verify client ID and client secret are correct in environment config
- Deploy refreshed tokens and confirm scoring requests succeed
Data Sync Failure
- Identify if scoring results are stale or inconsistent across requests
- Check if Grammarly updated their scoring model (review changelog)
- Compare current scores against known baseline text samples
- If quality drift confirmed: log evidence and file support ticket
- Consider pinning API version if Grammarly supports versioned endpoints
Communication Template
**Incident**: Grammarly Integration [Outage/Degradation] **Status**: [Investigating/Identified/Mitigating/Resolved] **Started**: YYYY-MM-DD HH:MM UTC **Impact**: [Text scoring unavailable / elevated latency / OAuth failure affecting N users] **Current action**: [Fallback scores active / token refresh in progress / rate limit backoff enabled] **Next update**: HH:MM UTC
Post-Incident
- Document timeline from detection to resolution
- Identify root cause (Grammarly outage / token expiry / rate limit exceeded)
- Verify all scoring pipelines resumed with accurate results
- Add proactive token refresh before expiry (buffer by 10 min)
- Implement rate limit monitoring with alerting at 80% threshold
- Update fallback logic if edge cases were discovered
Error Handling
| Incident Type | Detection | Resolution |
|---|---|---|
| Text check timeout | Requests exceeding 10s SLA | Enable timeout + retry with exponential backoff |
| Suggestion quality degradation | Scores deviate from baseline samples | Log evidence, pin API version, file support ticket |
| OAuth token failure | 401 on all authenticated endpoints | Trigger refresh flow, re-authorize if refresh fails |
| Rate limit storm | 429 responses with Retry-After header | Implement request queuing, reduce concurrency |
Resources
Next Steps
See
grammarly-observability for monitoring setup and alerting thresholds.