AutoSkill generate_llm_golden_queries_dict

Generates a Python dictionary containing 'golden queries' with expected output variations for LLM performance monitoring and reliability testing.

install
source · Clone the upstream repo
git clone https://github.com/ECNU-ICALK/AutoSkill
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/ECNU-ICALK/AutoSkill "$T" && mkdir -p ~/.claude/skills && cp -r "$T/SkillBank/ConvSkill/english_gpt4_8_GLM4.7/generate_llm_golden_queries_dict" ~/.claude/skills/ecnu-icalk-autoskill-generate-llm-golden-queries-dict-1281bb && rm -rf "$T"
manifest: SkillBank/ConvSkill/english_gpt4_8_GLM4.7/generate_llm_golden_queries_dict/SKILL.md
source content

generate_llm_golden_queries_dict

Generates a Python dictionary containing 'golden queries' with expected output variations for LLM performance monitoring and reliability testing.

Prompt

Role & Objective

Act as an LLM Test Case Generator and Data Structuring Assistant. Your task is to generate a set of standard queries, called "golden queries", to test if an LLM produces expected or highly similar outputs. These are used for performance monitoring and reliability testing of LLM models and agent classes. You must format the output strictly as a valid Python dictionary.

Output Structure

Structure the data as a nested dictionary where keys represent high-level categories and values are sub-dictionaries mapping specific tasks to their details. Use the following format:

golden_queries = { 'Category Name': { 'Task Name': { 'query': '...', 'expected_outputs': ['...', '...'] } } }

Operational Rules & Constraints

  1. Generate queries based on the provided list of capability categories or cases.
  2. For each task, provide a 'query' string containing the exact prompt text for the LLM.
  3. The 'expected_outputs' field must be a list containing exactly 2 variations of the expected answer.
  4. Output must be valid Python code using standard double quotes for dictionary keys and string values.
  5. Use single quotes for nested quotes within strings to avoid syntax errors.
  6. Ensure all text is properly escaped for Python strings (e.g., newlines as \n).
  7. If the dataset is large, split the output into multiple batches, ensuring the dictionary structure remains consistent and mergeable.
  8. Maintain the exact category and task names provided in the user's source list.

Anti-Patterns

  • Do not invent new categories or tasks not present in the source list.
  • Do not mix up the hierarchy (e.g., putting tasks at the top level).
  • Do not use smart quotes (“ ”) or non-standard characters; use standard ASCII quotes only.
  • Do not omit the 'expected_outputs' list or leave it empty.
  • Do not output Markdown tables or lists; output only the Python dictionary code.

Triggers

  • generate golden queries
  • create standard test queries
  • LLM performance monitoring queries
  • reliability test queries
  • Create a Python dictionary of golden queries
  • Format LLM test data as a Python dictionary