Claude-night-market justify
git clone https://github.com/athola/claude-night-market
T=$(mktemp -d) && git clone --depth=1 https://github.com/athola/claude-night-market "$T" && mkdir -p ~/.claude/skills && cp -r "$T/plugins/imbue/skills/justify" ~/.claude/skills/athola-claude-night-market-justify && rm -rf "$T"
plugins/imbue/skills/justify/SKILL.mdThe simplest change that fixes the problem is the safest change to merge. Adding code is easy. Removing the need for code is engineering.
Justify
The Additive Bias Problem
AI models are trained to be helpful, which creates a systematic bias toward adding code rather than fixing root causes:
| AI Default Behavior | Correct Behavior |
|---|---|
| Add a workaround | Fix the root cause |
| Modify test expectations | Fix the implementation |
| Create a new helper | Use an existing one |
| Add error handling | Prevent the error |
| Add a compatibility shim | Remove the old code |
| Wrap in try/catch | Fix the exception source |
This skill audits changes for these patterns and requires explicit justification for each.
When to Use
- After completing implementation work
- Before committing or creating PRs
- When reviewing your own changes for quality
- When scope-guard flags RED/YELLOW zone
Audit Protocol
Step 1: Gather the Delta
# Determine base branch base=$(git merge-base master HEAD 2>/dev/null \ || git merge-base main HEAD 2>/dev/null) # Get change statistics git diff "$base" --stat git diff "$base" --shortstat git diff "$base" --diff-filter=A --name-only # new files git diff "$base" --diff-filter=M --name-only # modified files git diff "$base" --diff-filter=D --name-only # deleted files
Step 2: Compute Additive Bias Score
Score each dimension 0-3 (0 = clean, 3 = high bias):
| Signal | Weight | How to Measure |
|---|---|---|
| Line ratio | 2x | |
| New files | 2x | Count of |
| Test logic changes | 3x | Test assertion/expectation diffs |
| New abstractions | 1x | New classes, functions, modules |
| Workaround patterns | 2x | Try/catch, if/else guards added |
Line Ratio Scoring:
| Ratio | Score | Interpretation |
|---|---|---|
| < 2:1 | 0 | Balanced change |
| 2:1 to 5:1 | 1 | Mildly additive |
| 5:1 to 10:1 | 2 | Additive bias likely |
| > 10:1 | 3 | Strong additive bias |
Aggregate Score:
bias_score = sum(signal_score * weight) / sum(weights)
| Aggregate | Zone | Action |
|---|---|---|
| 0.0 - 0.5 | GREEN | Proceed |
| 0.5 - 1.5 | YELLOW | Justify each signal |
| 1.5 - 2.5 | RED | Rethink approach |
| 2.5+ | STOP | Likely wrong approach |
Step 3: Iron Law Compliance Check
The Iron Law states: tests drive implementation, not the other way around. Check for violations:
# Find test files that were modified git diff "$base" --name-only | rg "test_|_test\.|spec\." \ || git diff "$base" --name-only | grep -E "test_|_test\.|spec\." # For each modified test file, check what changed git diff "$base" -- <test_file> | rg "^[-+].*assert|^[-+].*expect|^[-+].*should"
Violation patterns (test logic was tampered):
- Assertion values changed (expected output modified)
- Test cases removed or commented out
or@skip
added@pytest.mark.skip- Error expectations weakened (broad exception types)
- Mock return values changed to match new behavior
- Test renamed to no longer describe original behavior
Each violation requires explicit justification:
"I changed this test assertion because the requirement changed, not because my implementation couldn't meet the original requirement."
If the requirement didn't change, the test should not change. Fix the implementation instead.
Step 4: Minimal Intervention Analysis
For each changed file, answer:
-
Was this change necessary? Could the goal be achieved without touching this file?
-
Was this the minimal change? Could fewer lines achieve the same result?
-
Did this change add or remove complexity? New functions, classes, or control flow = added complexity that needs justification.
-
Is there a subtraction-first alternative? Could removing code fix the problem instead of adding code?
Step 5: Generate Justification Report
Output a structured report:
## Justification Report **Branch**: feature/xyz **Base**: master **Delta**: +N/-M lines, X files changed ### Additive Bias Score: X.X (ZONE) | Signal | Score | Detail | |--------|-------|--------| | Line ratio | N | +A/-D = R:1 | | New files | N | [list] | | Test changes | N | [list] | | New abstractions | N | [list] | | Workarounds | N | [list] | ### Iron Law Compliance: PASS/FAIL [List any test logic modifications with justification] ### Change-by-Change Justification #### file.py (+N/-M) - **What**: [description] - **Why**: [root cause this addresses] - **Alternatives considered**: [what else could work] - **Why this is minimal**: [why fewer changes won't work] #### test_file.py (+N/-M) - **What**: [description] - **Justification**: [why test logic changed, if it did] - **Iron Law status**: PASS/VIOLATION ### Risk Assessment | Factor | Rating | |--------|--------| | Lines changed | LOW/MED/HIGH | | Files touched | LOW/MED/HIGH | | Test modifications | NONE/JUSTIFIED/VIOLATION | | New abstractions | NONE/JUSTIFIED/UNNECESSARY | | Overall merge risk | LOW/MED/HIGH | ### Recommendations [List any changes that should be reconsidered, simpler alternatives, or unnecessary additions]
Decision Weights
When evaluating competing approaches, weight these factors:
| Factor | Weight | Rationale |
|---|---|---|
| Fewer lines changed | HIGH | Less risk, easier review |
| No new files | HIGH | No new maintenance burden |
| No test logic changes | HIGH | Iron Law compliance |
| Root cause fix | HIGH | Prevents recurrence |
| Removes code | BONUS | Reduces maintenance surface |
| Adds abstraction | PENALTY | Only justified at 3rd use |
| Adds error handling | NEUTRAL | Only at system boundaries |
The Subtraction Test: Before accepting any change, ask: "Could I achieve this by removing code instead of adding it?" If yes, prefer the subtractive approach.
Integration with Proof of Work
Justify extends proof-of-work with change-level accountability:
- proof-of-work: "Did it work?" (evidence)
- justify: "Was this the right way?" (reasoning)
Both are required before claiming work is complete. Run proof-of-work first, then justify.
Anti-Patterns to Flag
1. Test Mutation
Changing test expectations to match broken code. Fix: Revert the test change, fix the implementation.
2. Shotgun Addition
Adding code in many files for a single-concern fix. Fix: Find the single point of change.
3. Defensive Overengineering
Adding try/catch, null checks, or validation for scenarios that can't happen in practice. Fix: Trust internal code. Only validate at boundaries.
4. Premature Abstraction
Creating a helper/utility/base class for one use case. Fix: Inline the code. Abstract at the 3rd use.
5. Compatibility Shim
Adding backward-compatibility code instead of updating callers. Fix: Update callers directly. Delete dead paths.