Skills prompt-ab-lab

Design, log, compare, and score prompt experiments so users can systematically

install
source · Clone the upstream repo
git clone https://github.com/openclaw/skills
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/openclaw/skills "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/52yuanchangxing/prompt-ab-lab" ~/.claude/skills/openclaw-skills-prompt-ab-lab && rm -rf "$T"
OpenClaw · Install into ~/.openclaw/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/openclaw/skills "$T" && mkdir -p ~/.openclaw/skills && cp -r "$T/skills/52yuanchangxing/prompt-ab-lab" ~/.openclaw/skills/openclaw-skills-prompt-ab-lab && rm -rf "$T"
manifest: skills/52yuanchangxing/prompt-ab-lab/SKILL.md
source content

Prompt A/B Lab

Purpose

Design, log, compare, and score prompt experiments so users can systematically improve outputs instead of guessing.

Trigger phrases

  • 比较两个提示词
  • prompt ab test
  • 提示词实验
  • 哪个 prompt 更好
  • 建一个评测表

Ask for these inputs

  • prompt A and B
  • task
  • evaluation criteria
  • test set
  • weights if any

Workflow

  1. Define what success looks like before comparing prompts.
  2. Generate an evaluation rubric and structured test table.
  3. Log outputs per test case and compute weighted scores.
  4. Summarize tradeoffs instead of declaring a winner too early.
  5. Recommend the next experiment iteration.

Output contract

  • experiment plan
  • scored comparison table
  • rubric
  • next-iteration suggestions

Files in this skill

  • Script:
    {baseDir}/scripts/prompt_experiment_logger.py
  • Resource:
    {baseDir}/resources/eval_rubric.md

Operating rules

  • Be concrete and action-oriented.
  • Prefer preview / draft / simulation mode before destructive changes.
  • If information is missing, ask only for the minimum needed to proceed.
  • Never fabricate metrics, legal certainty, receipts, credentials, or evidence.
  • Keep assumptions explicit.

Suggested prompts

  • 比较两个提示词
  • prompt ab test
  • 提示词实验

Use of script and resources

Use the bundled script when it helps the user produce a structured file, manifest, CSV, or first-pass draft. Use the resource file as the default schema, checklist, or preset when the user does not provide one.

Boundaries

  • This skill supports planning, structuring, and first-pass artifacts.
  • It should not claim that files were modified, messages were sent, or legal/financial decisions were finalized unless the user actually performed those actions.

Compatibility notes

  • Directory-based AgentSkills/OpenClaw skill.
  • Runtime dependency declared through
    metadata.openclaw.requires
    .
  • Helper script is local and auditable:
    scripts/prompt_experiment_logger.py
    .
  • Bundled resource is local and referenced by the instructions:
    resources/eval_rubric.md
    .