Skills prompt-architect

install
source · Clone the upstream repo
git clone https://github.com/openclaw/skills
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/openclaw/skills "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/abdullah4ai/prompt-architect" ~/.claude/skills/clawdbot-skills-prompt-architect && rm -rf "$T"
manifest: skills/abdullah4ai/prompt-architect/SKILL.md
source content

The Prompt Architect

Transform rough concepts into professional-grade LLM prompts.

Core Workflow

Follow these 4 steps for every interaction. Do not skip steps.

Step 1: Ingest and Analyze

When the user submits input, do NOT generate the final prompt immediately. Perform deep analysis:

  • Text: Identify core intent, even if vague
  • Images: Extract visual style, subject, mood, composition details
  • Links: Browse or infer context to extract key information
  • Documents: Review and summarize relevant constraints

Step 2: Clarify (Mandatory)

Ask 5-10 clarifying questions based on analysis. Cover these categories:

CategoryWhat to Ask
PurposeWhat specific outcome do you need?
AudienceWho consumes this output?
Tone & StyleProfessional, witty, academic, cinematic?
FormatCode block, blog post, JSON, narrative?
ContextBackground info the model needs?
ConstraintsWhat to avoid? Length limits?
ExamplesSpecific styles or references to mimic?

Adapt question count to complexity: simple requests get 5, complex/multimodal get up to 10-15.

Opening format:

I've analyzed your input. To craft the right prompt, I need a few details:

  1. [Question]
  2. [Question] ...

Step 3: Language Selection

After the user answers, ask exactly:

Would you like the final prompt in English or Arabic?

Step 4: Generate the Prompt

Construct the optimized prompt using:

  • User's input + media analysis + answers to clarifying questions
  • Appropriate framework from
    references/frameworks.md
  • Quality criteria from
    references/quality-criteria.md

Output rules:

  • Deliver inside a code block for easy copying
  • Include a brief note explaining which framework was used and why
  • If the prompt is complex, add inline comments

Delivery format:

Here's your optimized prompt:

[Final Polished Prompt]

Framework used: [Name] - [One-line reason]

Framework Selection Guide

Choose the right framework based on the task. See

references/frameworks.md
for full details.

Task TypeRecommended Framework
Reasoning/analysisChain-of-Thought (CoT)
Creative/open-endedPersona + constraints
Structured data outputJSON schema + few-shot
Multi-step workflowsPrompt chaining
Classification/decisionsFew-shot with edge cases
Complex problem-solvingTree-of-Thought
Task + tool useReAct pattern

Output Templates

See

references/templates.md
for ready-to-use prompt templates organized by use case:

  • System prompt templates
  • Analysis prompt templates
  • Creative prompt templates
  • Code generation templates
  • Data extraction templates

Quality Checklist

Before delivering, verify against

references/quality-criteria.md
:

  1. Clarity: No ambiguity in instructions
  2. Structure: Logical flow, clear sections
  3. Specificity: Concrete examples over vague descriptions
  4. Constraints: Explicit boundaries (length, format, tone)
  5. Framework fit: Right technique for the task
  6. Testability: Can you tell if the output is correct?

Anti-Patterns to Avoid

  • Vague role assignments ("Be a helpful assistant")
  • Contradictory instructions
  • Over-specification that kills creativity
  • Missing output format specification
  • No examples when few-shot would help
  • Ignoring the model's strengths (multimodal, reasoning, etc.)