Awesome-omni-skill Adaptive Bitrate

Adaptive Bitrate (ABR) streaming automatically adjusts video quality

install
source · Clone the upstream repo
git clone https://github.com/diegosouzapw/awesome-omni-skill
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/diegosouzapw/awesome-omni-skill "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/development/adaptive-bitrate" ~/.claude/skills/diegosouzapw-awesome-omni-skill-adaptive-bitrate-5b48f2 && rm -rf "$T"
manifest: skills/development/adaptive-bitrate/SKILL.md
source content

Adaptive Bitrate

Skill Profile

(Select at least one profile to enable specific modules)

  • DevOps
  • Backend
  • Frontend
  • AI-RAG
  • Security Critical

Overview

Adaptive Bitrate (ABR) streaming automatically adjusts video quality based on network conditions. This guide covers HLS, DASH, and player implementation for building video streaming solutions that provide smooth playback across varying network conditions.

Why This Matters

Adaptive Bitrate streaming is critical for video platforms as it directly impacts:

  • User Experience: Smooth playback without buffering
  • Reach: Works across diverse network conditions globally
  • Quality: Delivers optimal quality based on available bandwidth
  • Efficiency: Reduces bandwidth costs through smart quality selection

Core Concepts

  1. ABR (Adaptive Bitrate): Dynamic quality adjustment based on network conditions
  2. HLS (HTTP Live Streaming): Apple's streaming protocol using m3u8 playlists
  3. DASH (MPEG-DASH): ISO standard for adaptive streaming
  4. Segmentation: Video divided into small chunks for flexible switching
  5. Manifest Files: Playlists/MPD describing available quality levels
  6. Bandwidth Estimation: Real-time network speed detection
  7. Buffer Management: Balancing quality and playback stability

Inputs / Outputs / Contracts

Skill Composition

  • Depends on: None
  • Compatible with: None
  • Conflicts with: None
  • Related Skills: None

Quick Start / Implementation Example

  1. Review requirements and constraints
  2. Set up development environment
  3. Implement core functionality following patterns
  4. Write tests for critical paths
  5. Run tests and fix issues
  6. Document any deviations or decisions
# Example implementation following best practices
def example_function():
    # Your implementation here
    pass

Assumptions

  • Source video is in compatible format (MP4, MOV, etc.)
  • Sufficient storage for multiple quality versions
  • CDN available for global distribution
  • Modern browser with Media Source Extensions support

Compatibility

ProtocolBrowser Support
HLSSafari (native), Chrome/Firefox (via hls.js)
DASHChrome/Firefox (via Shaka), Safari (limited)
ProgressiveAll browsers

Test Scenario Matrix (QA Strategy)

TypeFocus AreaRequired Scenarios / Mocks
UnitCore LogicMust cover primary logic and at least 3 edge/error cases. Target minimum 80% coverage
IntegrationDB / APIAll external API calls or database connections must be mocked during unit tests
E2EUser JourneyCritical user flows to test
PerformanceLatency / LoadBenchmark requirements
SecurityVuln / AuthSAST/DAST or dependency audit
FrontendUX / A11yAccessibility checklist (WCAG), Performance Budget (Lighthouse score)

Technical Guardrails & Security Threat Model

1. Security & Privacy (Threat Model)

  • Top Threats: Injection attacks, authentication bypass, data exposure
  • Data Handling: Sanitize all user inputs to prevent Injection attacks. Never log raw PII
  • Secrets Management: No hardcoded API keys. Use Env Vars/Secrets Manager
  • Authorization: Validate user permissions before state changes

2. Performance & Resources

  • Execution Efficiency: Consider time complexity for algorithms
  • Memory Management: Use streams/pagination for large data
  • Resource Cleanup: Close DB connections/file handlers in finally blocks

3. Architecture & Scalability

  • Design Pattern: Follow SOLID principles, use Dependency Injection
  • Modularity: Decouple logic from UI/Frameworks

4. Observability & Reliability

  • Logging Standards: Structured JSON, include trace IDs
    request_id
  • Metrics: Track
    error_rate
    ,
    latency
    ,
    queue_depth
  • Error Handling: Standardized error codes, no bare except
  • Observability Artifacts:
    • Log Fields: timestamp, level, message, request_id
    • Metrics: request_count, error_count, response_time
    • Dashboards/Alerts: High Error Rate > 5%

Agent Directives

  1. Always encode multiple quality levels
  2. Use 6-10 second segments for optimal adaptation
  3. Configure appropriate buffer sizes
  4. Implement error handling for network issues
  5. Monitor quality switches and buffering
  6. Use CDN for global delivery

Definition of Done (DoD) Checklist

  • Tests passed + coverage met
  • Lint/Typecheck passed
  • Logging/Metrics/Trace implemented
  • Security checks passed
  • Documentation/Changelog updated
  • Accessibility/Performance requirements met (if frontend)

Anti-patterns / Pitfalls

  • Don't: Log PII, catch-all exception, N+1 queries
  • ⚠️ Watch out for: Common symptoms and quick fixes
  • 💡 Instead: Use proper error handling, pagination, and logging

Reference Links & Examples

  • Internal documentation and examples
  • Official documentation and best practices
  • Community resources and discussions

Versioning & Changelog

  • Version: 1.0.0
  • Changelog:
    • 2026-02-22: Initial version with complete template structure