Goclaw burner-phone
Control Android devices via ADB with vision feedback. Use this to see the screen, take screenshots, analyze UI elements, and automate phone tasks.
install
source · Clone the upstream repo
git clone https://github.com/smallnest/goclaw
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/smallnest/goclaw "$T" && mkdir -p ~/.claude/skills && cp -r "$T/internal/builtin_skills/burner-phone" ~/.claude/skills/smallnest-goclaw-burner-phone && rm -rf "$T"
manifest:
internal/builtin_skills/burner-phone/SKILL.mdsource content
Burner Phone Control
Use this skill for ANY request involving phone screens or mobile app automation.
Vision Feedback Loop
ALWAYS follow this pattern:
-
Screenshot: Capture the current screen
bash(cmd="adb exec-out screencap -p > ./assets/screen.png") -
Analyze: Use vision model to understand the screen
bash(cmd="python3 ./scripts/vision_helper.py ./assets/screen.png \"Describe the screen and list coordinates (x,y) for interactable elements.\"") -
Act: Perform the action using exact coordinates from step 2
bash(cmd="adb shell input tap <x> <y>") -
Verify: Screenshot again to confirm the action worked
Available Commands
Tapping
bash(cmd="adb shell input tap <x> <y>")
Swiping
bash(cmd="adb shell input swipe <x1> <y1> <x2> <y2> <duration_ms>")
Typing Text
bash(cmd="adb shell input text 'your text here'")
Key Events
bash(cmd="adb shell input keyevent KEYCODE_HOME") bash(cmd="adb shell input keyevent KEYCODE_BACK") bash(cmd="adb shell input keyevent KEYCODE_ENTER")
Launch App
bash(cmd="adb shell am start -n com.package.name/.MainActivity")
Rules
- ALWAYS screenshot before acting - never guess coordinates
- ALWAYS use vision_helper.py to get coordinates
- Use coordinates provided by the vision tool EXACTLY
- All paths are relative to the skill root directory