install
source · Clone the upstream repo
git clone https://github.com/openclaw/skills
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/openclaw/skills "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/4ier/claw-use" ~/.claude/skills/clawdbot-skills-claw-use && rm -rf "$T"
manifest:
skills/4ier/claw-use/SKILL.mdsource content
Claw Use — Device Control for AI Agents
Give your AI agent eyes, hands, and a voice on real devices.
Claw Use is a protocol + skill for AI agents to control physical devices over HTTP. The
cu CLI provides a unified interface — the same commands work across any device that implements the Claw Use API.
Supported Devices
| Platform | Implementation | Status |
|---|---|---|
| Android | claw-use-android | ✅ Available |
| iOS | claw-use-ios | 🔮 Planned |
| Desktop | claw-use-desktop | 🔮 Planned |
Prerequisites
CLI installed (ships with claw-use-android, or install standalone)cu- At least one device running a Claw Use implementation
- Device and agent on the same network (or connected via Tailscale)
Setup
# Add a device with a friendly name cu add redmi 192.168.0.105 <token> cu add pixel 100.80.1.10 <token> # List devices cu devices # ▸ redmi 192.168.0.105 online v1.2.0 # pixel 100.80.1.10 offline # Switch default cu use pixel # Target a specific device cu -d redmi screenshot
Core API (all platforms)
Every Claw Use implementation exposes the same HTTP endpoints:
Perception — read the device
cu screen # UI tree (semantic: element text, bounds, state) cu screen -c # compact mode (interactive elements only) cu screenshot # visual capture (JPEG, configurable quality) cu notifications # system notifications cu status # device health dashboard
Action — control the device
cu tap <x> <y> # tap coordinates cu click <text> # tap by visible text (semantic click) cu type "text" # type text (CJK supported) cu swipe up|down|left|right cu scroll up|down|left|right cu back / cu home # system navigation cu launch <app> # open an application cu open <url> # open URL cu intent '<json>' # platform-specific intent (Android)
Audio
cu tts "hello" # speak through device speaker cu say "你好" # alias
Device State
cu wake # wake screen cu lock / cu unlock # lock/unlock (PIN required for unlock)
Workflow Patterns
Navigate and interact
cu launch org.telegram.messenger cu screen -c # see what's on screen cu click "Search" cu type "John" cu click "John, last seen recently" cu type "Hey!" cu click "Send"
Visual + semantic dual-channel
cu screen -c # semantic: what elements exist cu screenshot 50 720 /tmp/look.jpg # visual: what it actually looks like
Multi-device orchestration
cu -d phone1 launch com.whatsapp cu -d phone2 screenshot cu -d tablet open "https://example.com"
For Agent Developers
Claw Use is designed as a protocol, not just an app. To add support for a new platform:
- Implement the Claw Use HTTP API spec
- Expose endpoints on a configurable port (default: 7333)
- Support token auth via
headerX-Bridge-Token - Return JSON responses matching the documented schemas
The
cu CLI and this skill work automatically with any compliant implementation.
Tips
is the primary perception tool — compact mode filters noisecu screen -c
by text is more reliable thancu click
when text is visiblecu tap
when you need visual context the UI tree can't capturecu screenshot- Auto-unlock is transparent: locked devices auto-unlock before any command
- Combine with Tailscale for remote access from anywhere