Awesome-omni-skill define

Create well-defined issues (bugs, features, tasks) using Socratic questioning to eliminate ambiguity. Use when creating new Linear or Jira issues, when the user mentions Linear, Jira, or needs to define work items with proper acceptance criteria and relationships.

install
source · Clone the upstream repo
git clone https://github.com/diegosouzapw/awesome-omni-skill
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/diegosouzapw/awesome-omni-skill "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/cli-automation/define" ~/.claude/skills/diegosouzapw-awesome-omni-skill-define && rm -rf "$T"
manifest: skills/cli-automation/define/SKILL.md
safety · automated scan (low risk)
This is a pattern-based risk scan, not a security review. Our crawler flagged:
  • global npm install
Always read a skill's source content before installing. Patterns alone don't mean the skill is malicious — but they warrant attention.
source content
<objective> Guide the creation of precise, unambiguous issues through the Socratic method. Ask targeted questions to uncover edge cases, acceptance criteria, relationships, and constraints before creating the issue. A well-defined issue prevents thrashing and enables clear work execution. </objective>

<backend_detection> Read backend from mobius config (

mobius.config.yaml
or
~/.config/mobius/config.yaml
). Default to 'linear' if not specified.

# mobius.config.yaml
backend: linear  # or 'jira' or 'local'

The backend determines the output format for issue specifications. </backend_detection>

<local_mode> When

backend: local
is detected, skip all CLI calls and write issue specs directly to the project-local
.mobius/
directory.

<local_mode_detection> Check the config file for

backend: local
:

# mobius.config.yaml
backend: local

If the backend is

local
, the entire CLI-based creation flow is bypassed. The Socratic questioning workflow remains identical — only the final output step changes. </local_mode_detection>

<local_id_generation> Generate a sequential

LOC-{N}
identifier using
.mobius/issues/counter.json
.

The counter file tracks the next available ID. Use an atomic read-increment-write pattern:

  1. Read the counter file (create with
    {"nextTaskNumber": 1, "lastUpdated": "..."}
    if missing)
  2. Capture the current
    nextTaskNumber
    value
  3. Increment and write the updated counter back immediately
  4. Format the identifier as
    LOC-{N}
    zero-padded to 3 digits (e.g.,
    LOC-001
    ,
    LOC-012
    )
# Read current counter (or initialize if missing)
COUNTER_FILE=".mobius/issues/counter.json"

if [ ! -f "$COUNTER_FILE" ]; then
  mkdir -p .mobius/issues
  echo '{"nextTaskNumber": 1, "lastUpdated": "'$(date -u +%Y-%m-%dT%H:%M:%SZ)'"}' > "$COUNTER_FILE"
fi

# Read the current value
NEXT_NUM=$(cat "$COUNTER_FILE" | jq '.nextTaskNumber')

After reading, immediately increment and write back:

# Increment and write back atomically
echo "{\"nextTaskNumber\": $((NEXT_NUM + 1)), \"lastUpdated\": \"$(date -u +%Y-%m-%dT%H:%M:%SZ)\"}" > "$COUNTER_FILE"

Format the identifier:

LOC_ID=$(printf "LOC-%03d" $NEXT_NUM)

Edge case — counter.json missing or corrupted: Scan existing

LOC-*
directories under
.mobius/issues/
to find the highest number, then set
nextTaskNumber
to max + 1. </local_id_generation>

<local_issue_creation> After user approval, write the issue spec to

.mobius/issues/LOC-{N}/parent.json
.

The JSON schema follows the

ParentIssueContext
type from
src/types/context.ts
:

{
  "id": "LOC-001",
  "identifier": "LOC-001",
  "title": "Issue title from Socratic questioning",
  "description": "## Summary\n\nFull markdown description with acceptance criteria...",
  "gitBranchName": "feature/loc-001",
  "status": "Backlog",
  "labels": ["Feature"],
  "url": ""
}

Field mapping for local mode:

FieldValue
id
Same as
identifier
(e.g.,
LOC-001
)
identifier
Generated
LOC-{N}
ID
title
Issue title from investigation
description
Full markdown description with acceptance criteria
gitBranchName
feature/{loc-id}
lowercase (e.g.,
feature/loc-001
)
status
"Backlog"
(initial state)
labels
Array of label strings (e.g.,
["Bug"]
,
["Feature"]
)
url
Empty string
""
(no backend URL)

Directory creation and file write:

# Create the issue directory
mkdir -p ".mobius/issues/${LOC_ID}/tasks"
mkdir -p ".mobius/issues/${LOC_ID}/execution"

Then use the Write tool to create

parent.json
with the full issue spec.

Also create the context.json wrapper file:

{
  "parent": { ... },
  "subTasks": [],
  "metadata": {
    "fetchedAt": "2026-01-30T12:00:00Z",
    "updatedAt": "2026-01-30T12:00:00Z",
    "backend": "local"
  }
}

Ensure

.mobius/.gitignore
exists: If
.mobius/.gitignore
does not exist, create it with:

# Mobius local state - gitignored by default
*
!.gitignore

</local_issue_creation>

<local_approval_workflow> The approval workflow for local mode mirrors the CLI flow, but outputs to files instead of API calls.

After Socratic questioning is complete, present the issue for approval (same as CLI flow):

"Here is the issue I'll create:

Title: [title] ID: [LOC-{N}] (local) Type/Labels: [Bug/Feature/Improvement] Priority: [Urgent/High/Normal/Low] Description: [full description with acceptance criteria]

Ready to create this local issue?"

Use AskUserQuestion:

  • Create issue - Issue looks correct, write to local files
  • Make changes - I need to modify something
  • Add more context - I have additional information
  • Cancel - Don't create this issue

After writing the files:

"Local issue created successfully!

{LOC-ID}: {Issue Title} Location:

.mobius/issues/{LOC-ID}/parent.json

Would you like to:

  • Break this down into sub-tasks (
    /refine {LOC-ID}
    )
  • Define another issue" </local_approval_workflow>

<local_mode_constraints> What changes in local mode:

  • No CLI calls (no
    linear
    or
    acli
    commands)
  • No workspace context fetching (teams, labels, projects) — ask user directly
  • No relationship linking to backend issues (relationships tracked locally if needed)
  • Issue spec written to
    .mobius/issues/LOC-{N}/parent.json

What stays the same:

  • Full Socratic questioning workflow
  • Acceptance criteria requirements
  • Approval workflow before creation
  • Issue description structure and quality standards
  • All anti-patterns and success criteria checks </local_mode_constraints> </local_mode>

<backend_context> <linear> Linear Concepts:

  • States: Backlog, Todo, In Progress, Done, Canceled, Duplicate
  • Labels: Bug, Feature, Improvement (and custom labels)
  • Priority: 0 (No priority), 1 (Urgent), 2 (High), 3 (Normal), 4 (Low)
  • Relationships: blocks, blockedBy, relatedTo, duplicateOf
  • Hierarchy: Issues can have parent issues (sub-issues) </linear>
<jira> **Jira Configuration**: Requires `project_key` from jira config section:
# mobius.config.yaml
backend: jira
jira:
  base_url: https://yourcompany.atlassian.net
  project_key: PROJ

Jira Concepts:

  • Statuses: To Do, In Progress, Done (varies by workflow)
  • Issue Types: Bug, Story, Task, Epic, Sub-task
  • Priority: Highest, High, Medium, Low, Lowest
  • Links: blocks, is blocked by, relates to, duplicates
  • Hierarchy: Epics contain Stories/Tasks; Tasks can have Sub-tasks </jira>

</backend_context>

<context_gathering_config> Read workspace context from

mobius.config.yaml
defaults.

Before presenting options to the user, read team, project, and label defaults from the config file.

For Linear backend — read from config:

linear:
  team: Engineering
  project: My Project
  default_labels: [Bug, Feature, Improvement]

For Jira backend — read from config:

jira:
  base_url: https://yourcompany.atlassian.net
  project_key: PROJ
  default_labels: [bug, story, task]

Workflow:

  1. Detect backend from config file (default: linear)
  2. Read workspace defaults (team, project, labels) from config
  3. Present config-based options in AskUserQuestion dialogs
  4. If config values are missing, ask user directly as fallback

CLI availability check (run once at start):

# For Linear backend
command -v linearis >/dev/null 2>&1 || echo "linearis not found. Install via: npm install -g linearis"

# For Jira backend
command -v acli >/dev/null 2>&1 || echo "acli not found. See: https://developer.atlassian.com/cloud/acli/"

For searching related issues (when user mentions related work):

# Linear: search for related issues
linearis issues search "search terms"

# Jira: search for related issues
acli jira workitem search --jql "summary ~ 'search terms' AND project = PROJ"

This approach avoids runtime metadata fetching and uses config defaults instead. </context_gathering_config>

<quick_start> <initial_gate> CRITICAL - Run this BEFORE investigation

If user provides no context (just invoked the skill), use AskUserQuestion:

Question: "What kind of issue do you need to create?"

Options:

  1. Bug report - Something is broken or not working as expected
  2. Feature request - New capability or enhancement
  3. Task - General work item
  4. Improvement - Enhancement to existing functionality </initial_gate>
<workflow> 1. **Determine issue type** - Bug, feature, task, or improvement 2. **Identify team/project** - From context or ask user directly 3. **Gather core information** - Title, description, affected areas 4. **Investigate with Socratic questions** - Ask until no ambiguities remain 5. **Define acceptance criteria** - Verifiable outcomes 6. **Identify relationships** - What blocks this? What does this block? 7. **Set priority and metadata** - Priority, labels/issue type, project 8. **Present for approval** - Show complete issue before creating 9. **Create issue via CLI** - Create issue directly using CLI commands after approval </workflow> </quick_start>

<socratic_investigation> <purpose> Uncover hidden requirements, edge cases, and ambiguities through targeted questioning. Each question should reveal information that prevents incorrect implementation or scope creep. </purpose>

<question_categories> <bug_questions>

  • "What is the expected behavior vs actual behavior?"
  • "Can you reproduce this consistently? What are the exact steps?"
  • "Does this affect all users or specific scenarios?"
  • "What error messages or logs are shown?"
  • "When did this start happening? Any recent changes?"
  • "What is the impact - blocking users, data loss, degraded experience?" </bug_questions>

<feature_questions>

  • "Who is the primary user of this feature?"
  • "What problem does this solve? What's the current workaround?"
  • "What is the minimum viable version of this feature?"
  • "Are there related features this should integrate with?"
  • "What should happen at the edges - empty input, maximum values, errors?"
  • "How will we know this feature is successful?" </feature_questions>

<task_questions>

  • "What is the definition of done for this task?"
  • "What existing code or systems does this touch?"
  • "Are there dependencies that must complete first?"
  • "Who needs to be informed when this is complete?"
  • "What could block progress on this?" </task_questions>

<universal_questions>

  • "What priority does this have? 1 (Urgent) to 4 (Low)?"
  • "Is there a deadline or due date?"
  • "Who should own this work?"
  • "Are there labels that should be applied?"
  • "Does this relate to any existing issues?"
  • "Should this be part of a project?"
  • "How will we verify each criterion is met?" (test command, manual step, or observable outcome) </universal_questions> </question_categories>

<questioning_protocol> Thorough Socratic Questioning for Zero Ambiguity

The goal is to create issues with no ambiguity for execution and verification.

Interactive questioning flow:

  1. Initial classification - Use AskUserQuestion for issue type
  2. Scope definition - Ask about affected files/systems
  3. Edge case exploration - "What should happen if X fails?"
  4. Acceptance validation - "How will we verify this is done?"
  5. Boundary confirmation - "What is explicitly NOT in scope?"
  6. Priority assessment - "How critical is this work?"
  7. Dependency mapping - "What blocks this or is blocked by it?"

Anti-Ambiguity Checklist (verify before creating):

  • No vague terms like "should work better" or "improve performance"
  • Each criterion has explicit verification method
  • Edge cases are documented with expected behavior
  • Scope is bounded (what's in AND what's out)
  • Dependencies are identified and linked

AskUserQuestion patterns:

For scope clarification:

Question: "Which parts of the system does this affect?"
Options:
1. **Frontend only** - UI components, styling, client-side logic
2. **Backend only** - API, database, server-side logic
3. **Full stack** - Both frontend and backend changes
4. **Infrastructure** - CI/CD, deployment, configuration

For edge case handling:

Question: "What should happen when the operation fails?"
Options:
1. **Show error message** - Display user-friendly error and allow retry
2. **Silent fallback** - Use default behavior without notification
3. **Block operation** - Prevent action until issue resolved
4. **Log and continue** - Record error but proceed with degraded functionality

For verification method:

Question: "How should we verify this criterion is met?"
Options:
1. **Automated test** - Unit/integration test that can run in CI
2. **Manual testing** - Step-by-step verification by human
3. **Observable behavior** - Visible in logs, metrics, or UI
4. **Code review** - Verified by inspecting the implementation

Continue questioning until all aspects are crystal clear. </questioning_protocol>

<latent_error_prevention> Watch for these common sources of unclear issues:

  • Assumed context: "The usual flow" - which flow exactly?
  • Implicit scope: "Handle errors" - which errors? How?
  • Missing acceptance criteria: "Should work better" - how do we verify?
  • Hidden dependencies: "After the API is ready" - which issue?
  • Vague priority: "Soon" - Urgent or Low? </latent_error_prevention> </socratic_investigation>

<issue_structure> <description_template> A well-structured description includes:

## Summary
[1-2 sentence overview of the issue]

## Current Behavior (bugs only)
[What happens now that shouldn't]

## Expected Behavior
[What should happen instead]

## Reproduction Steps (bugs only)
1. Step one
2. Step two
3. Observe issue

## Acceptance Criteria
- [ ] Criterion 1 with verifiable outcome
  - **Verification**: `test command` | manual step | observable
- [ ] Criterion 2 with test method
  - **Verification**: `test command` | manual step | observable
- [ ] Criterion 3 with manual verification step
  - **Verification**: `test command` | manual step | observable

## Additional Context
[Screenshots, logs, related issues]

</description_template>

<acceptance_criteria_rules> Write criteria as behavioral outcomes, not implementation details:

GOOD (outcomes):

  • "User can deactivate schedule without error"
  • "All team members see updated schedule within 30 seconds"
  • "Error message displays with actionable guidance"

BAD (implementation):

  • "Add try/catch around database call"
  • "Use WebSocket for real-time sync"
  • "Call scheduleService.deactivate()"

Each criterion should be:

  1. Observable - Can be seen or measured
  2. Verifiable - Has a test or check method
  3. Unambiguous - Only one interpretation </acceptance_criteria_rules> </issue_structure>

<checkpoint_issues> <purpose> Checkpoint issues represent decision points where human input is required before work can proceed. They're used when the implementation approach is genuinely ambiguous and an agent cannot make the decision autonomously. </purpose>

<when_to_create_checkpoints> Create a checkpoint issue when:

  • Technology choice - Multiple valid libraries/frameworks could solve the problem
  • Architecture decision - The approach has long-term implications
  • Business logic ambiguity - The "right" behavior depends on product decisions
  • Trade-off evaluation - Options have meaningfully different pros/cons

Do NOT create checkpoints for:

  • Implementation details an agent can reasonably decide
  • Standard patterns with clear best practices
  • Decisions that can be easily changed later
  • Personal style preferences </when_to_create_checkpoints>

<checkpoint_template>

## Summary
[Brief description of the decision that needs to be made]

## Type: checkpoint:decision

## Decision Required
[Specific question that needs answering - one decision per checkpoint]

## Options
1. **Option A** - [Name]
   - Pros: [Benefits of this approach]
   - Cons: [Drawbacks or trade-offs]
   - Example: [Code snippet or reference if helpful]

2. **Option B** - [Name]
   - Pros: [Benefits of this approach]
   - Cons: [Drawbacks or trade-offs]
   - Example: [Code snippet or reference if helpful]

3. **Option C** - [Name] (optional)
   - Pros: [Benefits of this approach]
   - Cons: [Drawbacks or trade-offs]
   - Example: [Code snippet or reference if helpful]

## Recommendation
[If the agent has a recommendation, state it with reasoning. Otherwise: "No strong recommendation - depends on team preference."]

## Default Behavior
If no decision is made within [timeframe, e.g., "24 hours" or "before next sprint"], proceed with: **[Option X]**
Reason: [Why this is a safe default]

## Blocks
This decision blocks: [List of dependent sub-tasks or issues]

</checkpoint_template>

<checkpoint_example> Scenario: Implementing dark mode feature requires deciding how to persist theme preference.

## Summary
Choose the approach for persisting user theme preference.

## Type: checkpoint:decision

## Decision Required
How should we persist the user's theme preference across sessions?

## Options
1. **localStorage only**
   - Pros: Simple implementation, no server changes, immediate read
   - Cons: Not available during SSR, can flash wrong theme on load
   - Example: `localStorage.setItem('theme', 'dark')`

2. **Cookie only**
   - Pros: Available during SSR, no theme flash
   - Cons: Sent with every request, 4KB limit, requires cookie parsing
   - Example: `document.cookie = 'theme=dark; max-age=31536000'`

3. **Cookie + localStorage hybrid**
   - Pros: SSR-friendly AND fast client reads, best UX
   - Cons: More complex, must keep in sync
   - Example: Cookie for SSR, localStorage for client preference changes

## Recommendation
**Option 3 (hybrid)** if SSR is used, otherwise **Option 1 (localStorage)**.
Our app uses Next.js with SSR, so the hybrid approach prevents theme flash.

## Default Behavior
If no decision is made within 24 hours, proceed with: **Option 1 (localStorage)**
Reason: Simplest implementation; theme flash is acceptable for initial release.

## Blocks
This decision blocks: MOB-125 (Create ThemeProvider), MOB-126 (Add useTheme hook)

</checkpoint_example>

<timeout_handling> Every checkpoint must have a default behavior to prevent indefinite blocking:

  1. Specify a timeout - Usually 24-48 hours or "before next sprint"
  2. Choose a safe default - The option that is easiest to change later
  3. Explain the reasoning - Why this default won't cause problems

If the checkpoint is critical and has no safe default, escalate to the issue creator. </timeout_handling>

<checkpoint_resolution> When a decision is made:

  1. Add a comment with the decision and reasoning
  2. Update the issue description with "Decision: Option X selected"
  3. Move the checkpoint to Done
  4. Unblock dependent issues

The agent executing dependent tasks should read the checkpoint's decision before implementing. </checkpoint_resolution> </checkpoint_issues>

<context_gathering> <gathering_approach> Read workspace context from

mobius.config.yaml
, then present options to the user.

Gathering workflow:

  1. Read
    team
    ,
    project
    , and
    default_labels
    from config
    linear:
    section
  2. Present team/project as defaults via AskUserQuestion (user can override)
  3. Present label options from
    default_labels
    via AskUserQuestion
  4. If config values are missing, ask user directly

For Jira:

  1. Read
    project_key
    and
    default_labels
    from config
    jira:
    section
  2. Present issue type options from
    default_labels
    via AskUserQuestion
  3. If config values are missing, ask user directly

For related issues: Ask the user if they know of related issues, or use CLI to search:

# Linear: search for related issues
linearis issues search "search terms"

# Jira: search for related issues
acli jira workitem search --jql "summary ~ 'search terms' AND project = PROJ"

</gathering_approach>

<relationship_discovery> Ask about relationships:

  • "Do you know of any existing issues this is related to?"
  • "Should any existing issues be blocked by this work?"
  • "Is this a duplicate of an existing issue?"

Linear relationships (included in output):

  • blocks
    : Issues this one blocks
  • blockedBy
    : Issues blocking this one
  • relatedTo
    : Related issues
  • duplicateOf
    : If this duplicates another issue

Jira links (included in output):

  • blocks / is blocked by
  • relates to
  • duplicates / is duplicated by </relationship_discovery> </context_gathering>

<priority_guidelines> <priority_matrix>

PriorityLinearJiraWhen to use
Urgent1HighestProduction down, data loss, security
High2HighMajor feature broken, many users affected
Normal3MediumImportant but not urgent, default
Low4LowNice to have, improvements
None0LowestNot yet triaged
</priority_matrix>

<priority_questions> Use AskUserQuestion for priority:

Question: "What priority should this have?"

Options:

  1. Urgent - Production impact, must fix immediately
  2. High - Major functionality affected, fix soon
  3. Normal - Important but not urgent (recommended default)
  4. Low - Enhancement, can wait </priority_questions> </priority_guidelines>

<approval_workflow> <before_output> Present the complete issue in chat:

"Here is the issue I'll create:

Title: [title] Team/Project: [team or project name] Type/Labels: [Bug/Feature/Improvement or issue type] Priority: [Urgent/High/Normal/Low] State/Status: [initial state] Description: [full description with acceptance criteria]

Relationships: [if any] Parent: [if applicable]

Ready to create this issue?"

Use AskUserQuestion:

  • Create issue - Issue looks correct, create it
  • Make changes - I need to modify something
  • Add more context - I have additional information
  • Cancel - Don't create this issue </before_output>

<cli_creation> After approval, create the issue using CLI commands via Bash.

For Linear — use

linearis issues create
:

linearis issues create "{issue title}" \
  --team "{team from config}" \
  --description "{full description with acceptance criteria}" \
  --priority {1-4} \
  --state "{initial state}" \
  --labels "{label1},{label2}"

For Jira — use

acli jira workitem create
:

acli jira workitem create \
  --project "{project_key from config}" \
  --type "{Bug|Story|Task}" \
  --summary "{issue title}" \
  --description "{full description with acceptance criteria}" \
  --priority "{High|Medium|Low}"

After creation, parse the JSON output to extract the issue ID and URL.

If CLI is not installed, report a clear error:

<after_creation> After creating the issue via CLI:

"Issue created successfully!

{Issue ID}: {Issue Title} URL: {issue URL from CLI output}

Would you like to:

  • Define related issues
  • Break this down into sub-tasks (/refine)
  • Add this to a project/epic" </after_creation> </approval_workflow>
<examples> <bug_example> User: "There's a bug with schedules"

Response flow with AskUserQuestion:

  1. Initial question: "What is happening vs what should happen?"
  2. Follow-up: "Can you reproduce this? What are the exact steps?"
  3. Scope question via AskUserQuestion:
    • Question: "Does this affect all users or specific scenarios?"
    • Options: All users, Specific user roles, Specific data conditions, Unknown
  4. Priority question via AskUserQuestion:
    • Question: "What is the impact of this bug?"
    • Options: Production down, Major feature broken, Degraded experience, Minor annoyance

After gathering all details, present for approval:

"Here is the issue I'll create:

Title: Schedule deactivation throws 500 error Team: Engineering Type: Bug Priority: Urgent (1) State: Todo

Description:

Summary

Users receive HTTP 500 error when deactivating schedules. ...

Ready to create this issue?"

After approval, create via CLI:

linearis issues create "Schedule deactivation throws 500 error" \
  --team "Engineering" \
  --description "## Summary
Users receive HTTP 500 error when deactivating schedules.

## Current Behavior
Clicking 'Deactivate' shows error toast and schedule remains active.

## Expected Behavior
Schedule deactivates successfully with confirmation message.

## Reproduction Steps
1. Navigate to Schedule Settings
2. Click 'Deactivate Schedule'
3. Observe 500 error in toast

## Acceptance Criteria
- [ ] User can deactivate schedule without error
  - **Verification**: Manual test - click Deactivate, observe success toast
- [ ] Schedule status updates to 'inactive'
  - **Verification**: \`npm test -- --grep 'schedule deactivation'\`
- [ ] Team members see schedule status change
  - **Verification**: Manual test - check team view after deactivation
- [ ] Error logs capture root cause for monitoring
  - **Verification**: Observable - check logs after fix deployment" \
  --priority 1 \
  --state "Todo" \
  --labels "Bug"

Report result to user:

"Issue created successfully!

MOB-200: Schedule deactivation throws 500 error URL: https://linear.app/mobius/issue/MOB-200

Would you like to break this down into sub-tasks (/refine)?" </bug_example>

<feature_example> User: "We need to add dark mode"

Response flow with thorough AskUserQuestion:

  1. "Who is the primary user of this feature?"
  2. AskUserQuestion for behavior:
    • Question: "How should the theme be controlled?"
    • Options: Follow system only, Manual toggle only, Both system and manual, User choice on first launch
  3. AskUserQuestion for scope:
    • Question: "Which areas need dark mode support?"
    • Options: All screens, Core screens only, Settings and main views, Specific components (list them)
  4. AskUserQuestion for verification:
    • Question: "How should we verify the feature works?"
    • Options: Automated visual regression, Manual QA checklist, Accessibility audit, All of the above

After gathering all details and approval, create via CLI:

linearis issues create "Add dark mode theme support" \
  --team "Engineering" \
  --description "## Summary
Add dark mode support with system preference detection and manual toggle.

## Expected Behavior
- App detects system dark mode preference on launch
- User can manually toggle between light/dark/system
- All screens render correctly in both modes

## Scope
**In scope**: All core screens, settings, navigation
**Out of scope**: Admin dashboard (separate issue)

## Acceptance Criteria
- [ ] Theme follows system preference by default
  - **Verification**: \`npm test -- --grep 'theme system preference'\`
- [ ] Settings screen has theme toggle (Light/Dark/System)
  - **Verification**: Manual test - navigate to Settings, verify toggle exists
- [ ] All text maintains 4.5:1 contrast ratio in both modes
  - **Verification**: \`npm run test:a11y\` or Lighthouse accessibility audit
- [ ] Theme preference persists across app restarts
  - **Verification**: Manual test - set theme, restart app, verify theme persists
- [ ] No flash of wrong theme on app launch
  - **Verification**: Observable - launch app in dark mode, no white flash

## Edge Cases
- If localStorage is unavailable, default to system preference
- If system preference API unavailable, default to light mode" \
  --priority 3 \
  --state "Backlog" \
  --labels "Feature"

Report result to user:

"Issue created successfully!

MOB-201: Add dark mode theme support URL: https://linear.app/mobius/issue/MOB-201

This is a larger feature. Would you like to break it down into sub-tasks (/refine MOB-201)?" </feature_example> </examples>

<issue_creation_cli> Create the issue using CLI commands via Bash after user approval.

This skill creates issues directly via CLI — no structured YAML output needed.

For Linear — use

linearis issues create
:

# Write description to a temp file for multi-line content
DESCRIPTION=$(cat <<'DESC'
## Summary
{description content}

## Acceptance Criteria
- [ ] Criterion 1
  - **Verification**: test command or manual step
DESC
)

linearis issues create "{issue title}" \
  --team "{team from config}" \
  --description "$DESCRIPTION" \
  --priority {1-4} \
  --state "{initial state}" \
  --labels "{label1},{label2}"

For Jira — use

acli jira workitem create
:

DESCRIPTION=$(cat <<'DESC'
## Summary
{description content}

## Acceptance Criteria
- [ ] Criterion 1
  - **Verification**: test command or manual step
DESC
)

acli jira workitem create \
  --project "{project_key from config}" \
  --type "{Bug|Story|Task}" \
  --summary "{issue title}" \
  --description "$DESCRIPTION" \
  --priority "{High|Medium|Low}"

After successful creation:

  1. Parse the JSON output to extract the issue ID and URL
  2. Report the created issue to the user
  3. Offer next steps (refine, create related issues)

If CLI command fails:

  1. Report the error to the user
  2. If CLI not found, show install instructions:
  3. If auth error, suggest checking CLI authentication
  4. Offer to retry or save issue details for manual creation

Relationship handling:

  • blocks
    : Issues that cannot start until this one completes
  • blockedBy
    : Issues that must complete before this one can start
  • relatedTo
    : Related issues for reference
  • duplicateOf
    : Mark as duplicate of existing issue

Backend-specific field mappings:

FieldLinear (CLI)Jira (acli)
team
/
project
--team
flag
--project
flag
priority
--priority {1-4}
--priority {High/Medium/Low}
state
/
status
--state
flag
N/A (uses workflow default)
labels
/
type
--labels
flag (comma-separated)
--type
flag
</issue_creation_cli>

<anti_patterns> Don't accept vague requirements:

  • BAD: "Fix the scheduling bug"
  • GOOD: "Fix 500 error on schedule deactivation affecting all users"

Don't skip acceptance criteria:

  • BAD: Create issue with just title and description
  • GOOD: Every issue has verifiable acceptance criteria

Don't assume priority:

  • BAD: Default everything to Normal
  • GOOD: Ask about impact and urgency to determine priority

Don't ignore relationships:

  • BAD: Create isolated issues
  • GOOD: Search for related issues and set up relationship links

Don't create compound issues:

  • BAD: "Fix deactivation and add team sync and improve UI"
  • GOOD: Create separate issues for each concern

Don't write untestable acceptance criteria:

  • BAD: "System should be faster" (unmeasurable)
  • BAD: "UI should look better" (subjective)
  • GOOD: "Page load time < 2 seconds" with
    Verification: Lighthouse performance score > 90
  • GOOD: "Button uses primary color from design system" with
    Verification: Visual regression test
    </anti_patterns>

<success_criteria> An issue is ready when:

  • Type/labels match the nature of the work (Bug/Feature/Task)
  • Title is specific and actionable
  • Description includes all relevant context
  • Acceptance criteria are behavioral outcomes
  • Each criterion has explicit verification method
  • Priority reflects actual urgency/impact
  • Relationships are identified and linked
  • Project/team is set
  • No vague terms remain (no "should work better", "improve performance")
  • Edge cases are documented with expected behavior
  • Scope is bounded (what's in AND what's out)
  • User has approved before creating
  • Issue created successfully via CLI command
  • Created issue URL returned to user </success_criteria>