Openpaw c-ai
Query LLMs from the CLI — pipe text for summarization, chat interactively, use local or cloud models with llm or aichat.
install
source · Clone the upstream repo
git clone https://github.com/daxaur/openpaw
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/daxaur/openpaw "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/c-ai" ~/.claude/skills/daxaur-openpaw-c-ai && rm -rf "$T"
manifest:
skills/c-ai/SKILL.mdsource content
AI / LLM Tools
llm (Simon Willison)
# Quick prompt llm "What is the capital of France?" # Pipe text for processing cat article.txt | llm "Summarize this in 3 bullet points" git diff | llm "Write a commit message for these changes" pbpaste | llm "Fix the grammar in this text" # Interactive chat llm chat # Use specific model llm -m claude-3.5-sonnet "Explain quantum computing" llm -m gpt-4o "Review this code" # List available models llm models # Install model plugins llm install llm-claude-3 llm install llm-ollama # local models # View prompt/response history llm logs list llm logs last
aichat
# Quick prompt aichat "Explain Docker in simple terms" # Pipe input cat code.py | aichat "Find bugs in this code" # Interactive REPL aichat # Shell assistant (generates and runs commands) aichat -e "find all files larger than 100MB" # Specific model aichat -m claude-3.5-sonnet "Hello" # List models aichat --list-models
Guidelines
- Use
for piping text through LLMs (summarize, translate, analyze)llm - Use
for generating shell commands from natural languageaichat -e - Both tools store API keys locally — set up once with auth commands
has the richest plugin ecosystem (100+ model providers)llm
is faster (Rust) and has built-in RAG supportaichat- These tools use separate API keys from Claude Code — user pays per token