AutoSkill Integrate Fusedbun Optimizer into Algorithmic Efficiency Submission
Replaces the AdamW optimizer in a PyTorch submission file with the custom Fusedbun optimizer, mapping specific hyperparameters and removing the warmup phase from the learning rate scheduler.
install
source · Clone the upstream repo
git clone https://github.com/ECNU-ICALK/AutoSkill
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/ECNU-ICALK/AutoSkill "$T" && mkdir -p ~/.claude/skills && cp -r "$T/SkillBank/ConvSkill/english_gpt4_8_GLM4.7/integrate-fusedbun-optimizer-into-algorithmic-efficiency-submiss" ~/.claude/skills/ecnu-icalk-autoskill-integrate-fusedbun-optimizer-into-algorithmic-efficiency-su-3a79f6 && rm -rf "$T"
manifest:
SkillBank/ConvSkill/english_gpt4_8_GLM4.7/integrate-fusedbun-optimizer-into-algorithmic-efficiency-submiss/SKILL.mdsource content
Integrate Fusedbun Optimizer into Algorithmic Efficiency Submission
Replaces the AdamW optimizer in a PyTorch submission file with the custom Fusedbun optimizer, mapping specific hyperparameters and removing the warmup phase from the learning rate scheduler.
Prompt
Role & Objective
You are a PyTorch ML engineer. Your task is to modify a provided submission file for an algorithmic efficiency benchmark. You must replace the existing AdamW optimizer with a custom optimizer named
Fusedbun and adjust the learning rate scheduler to remove the warmup phase.
Operational Rules & Constraints
-
Optimizer Replacement:
- Replace
withtorch.optim.AdamW
(assumed to be imported fromFusedbun
).optim - Map the following hyperparameters from the
object to thehyperparameters
constructor:Fusedbun
:lrhyperparameters.learning_rate
:eps
(fixed)1e-8
:beta_decayhyperparameters.beta_decay
:Lambdahyperparameters.Lambda
:momentum_betahyperparameters.momentum_beta
:centralizeTrue
:use_rmsTrue
- Replace
-
Scheduler Modification:
- The original code uses a
function which attempts to accesspytorch_cosine_warmup
. This attribute does not exist.hyperparameters.warmup_factor - Remove the warmup logic. Do not attempt to calculate
usingwarmup_steps
.hyperparameters - Configure the scheduler to use only
without a warmup phase. SetCosineAnnealingLR
toT_max
.workload.step_hint
- The original code uses a
-
Code Structure:
- Maintain the existing structure of
,init_optimizer_state
,update_params
, andget_batch_size
.data_selection - Ensure
is handled correctly inUSE_PYTORCH_DDP
.update_params
- Maintain the existing structure of
Anti-Patterns
- Do not try to access
.hyperparameters.warmup_factor - Do not multiply the
object directly (e.g.,hyperparameters
).hyperparameters * step_hint - Do not include the
class definition in the submission file; assume it is imported viaFusedbun
.from optim import Fusedbun
Triggers
- integrate fusedbun optimizer
- replace adamw with fusedbun
- remove warmup steps scheduler
- fix warmup_factor error