Babysitter Edge Deployment Skill
ML model optimization and deployment on robot edge devices (Jetson, embedded)
install
source · Clone the upstream repo
git clone https://github.com/a5c-ai/babysitter
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/a5c-ai/babysitter "$T" && mkdir -p ~/.claude/skills && cp -r "$T/library/specializations/robotics-simulation/skills/edge-deployment" ~/.claude/skills/a5c-ai-babysitter-edge-deployment-skill && rm -rf "$T"
manifest:
library/specializations/robotics-simulation/skills/edge-deployment/SKILL.mdsource content
Edge Deployment Skill
Overview
Expert skill for optimizing and deploying machine learning models on robot edge devices including NVIDIA Jetson and embedded systems.
Capabilities
- Configure TensorRT optimization for NVIDIA Jetson
- Set up ONNX model conversion and optimization
- Implement INT8 and FP16 quantization
- Configure DeepStream for video analytics
- Set up CUDA graph optimization
- Implement model pruning and distillation
- Configure DLA (Deep Learning Accelerator) deployment
- Set up multi-stream inference
- Implement ROS2 inference nodes
- Profile and benchmark on target hardware
Target Processes
- nn-model-optimization.js
- object-detection-pipeline.js
- rl-robot-control.js
- field-testing-validation.js
Dependencies
- TensorRT
- ONNX Runtime
- NVIDIA Jetson SDK
- DeepStream
Usage Context
This skill is invoked when processes require deploying ML models on edge devices with optimized inference performance.
Output Artifacts
- TensorRT engine files
- ONNX optimized models
- Quantization configurations
- DeepStream pipeline configs
- Inference benchmark reports
- ROS2 inference node implementations