Codex-plugin-cc gpt-5-4-prompting
Internal guidance for composing Codex and GPT-5.4 prompts for coding, review, diagnosis, and research tasks inside the Codex Claude Code plugin
install
source · Clone the upstream repo
git clone https://github.com/openai/codex-plugin-cc
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/openai/codex-plugin-cc "$T" && mkdir -p ~/.claude/skills && cp -r "$T/plugins/codex/skills/gpt-5-4-prompting" ~/.claude/skills/openai-codex-plugin-cc-gpt-5-4-prompting && rm -rf "$T"
manifest:
plugins/codex/skills/gpt-5-4-prompting/SKILL.mdsource content
GPT-5.4 Prompting
Use this skill when
codex:codex-rescue needs to ask Codex or another GPT-5.4-based workflow for help.
Prompt Codex like an operator, not a collaborator. Keep prompts compact and block-structured with XML tags. State the task, the output contract, the follow-through defaults, and the small set of extra constraints that matter.
Core rules:
- Prefer one clear task per Codex run. Split unrelated asks into separate runs.
- Tell Codex what done looks like. Do not assume it will infer the desired end state.
- Add explicit grounding and verification rules for any task where unsupported guesses would hurt quality.
- Prefer better prompt contracts over raising reasoning or adding long natural-language explanations.
- Use XML tags consistently so the prompt has stable internal structure.
Default prompt recipe:
: the concrete job and the relevant repository or failure context.<task>
or<structured_output_contract>
: exact shape, ordering, and brevity requirements.<compact_output_contract>
: what Codex should do by default instead of asking routine questions.<default_follow_through_policy>
or<verification_loop>
: required for debugging, implementation, or risky fixes.<completeness_contract>
or<grounding_rules>
: required for review, research, or anything that could drift into unsupported claims.<citation_rules>
When to add blocks:
- Coding or debugging: add
,completeness_contract
, andverification_loop
.missing_context_gating - Review or adversarial review: add
,grounding_rules
, andstructured_output_contract
.dig_deeper_nudge - Research or recommendation tasks: add
andresearch_mode
.citation_rules - Write-capable tasks: add
so Codex stays narrow and avoids unrelated refactors.action_safety
How to choose prompt shape:
- Use built-in
orreview
commands when the job is reviewing local git changes. Those prompts already carry the review contract.adversarial-review - Use
when the task is diagnosis, planning, research, or implementation and you need to control the prompt more directly.task - Use
for follow-up instructions on the same Codex thread. Send only the delta instruction instead of restating the whole prompt unless the direction changed materially.task --resume-last
Working rules:
- Prefer explicit prompt contracts over vague nudges.
- Use stable XML tag names that match the block names from the reference file.
- Do not raise reasoning or complexity first. Tighten the prompt and verification rules before escalating.
- Ask Codex for brief, outcome-based progress updates only when the task is long-running or tool-heavy.
- Keep claims anchored to observed evidence. If something is a hypothesis, say so.
Prompt assembly checklist:
- Define the exact task and scope in
.<task> - Choose the smallest output contract that still makes the answer easy to use.
- Decide whether Codex should keep going by default or stop for missing high-risk details.
- Add verification, grounding, and safety tags only where the task needs them.
- Remove redundant instructions before sending the prompt.
Reusable blocks live in references/prompt-blocks.md. Concrete end-to-end templates live in references/codex-prompt-recipes.md. Common failure modes to avoid live in references/codex-prompt-antipatterns.md.