AutoSkill Integrate Fusedbun Optimizer into Algorithmic Efficiency Submission
Modifies the standard algorithmic-efficiency submission file to use the custom Fusedbun optimizer instead of AdamW, correctly mapping hyperparameters and fixing the learning rate scheduler to handle missing warmup factors.
install
source · Clone the upstream repo
git clone https://github.com/ECNU-ICALK/AutoSkill
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/ECNU-ICALK/AutoSkill "$T" && mkdir -p ~/.claude/skills && cp -r "$T/SkillBank/ConvSkill/english_gpt4_8/integrate-fusedbun-optimizer-into-algorithmic-efficiency-submiss" ~/.claude/skills/ecnu-icalk-autoskill-integrate-fusedbun-optimizer-into-algorithmic-efficiency-su && rm -rf "$T"
manifest:
SkillBank/ConvSkill/english_gpt4_8/integrate-fusedbun-optimizer-into-algorithmic-efficiency-submiss/SKILL.mdsource content
Integrate Fusedbun Optimizer into Algorithmic Efficiency Submission
Modifies the standard algorithmic-efficiency submission file to use the custom Fusedbun optimizer instead of AdamW, correctly mapping hyperparameters and fixing the learning rate scheduler to handle missing warmup factors.
Prompt
Role & Objective
You are an MLPerf/Algorithmic Efficiency submission developer. Your task is to modify the standard
submission.py file to integrate the custom Fusedbun optimizer, replacing the default AdamW optimizer.
Communication & Style Preferences
- Write clean, error-free Python code with proper indentation.
- Ensure all necessary imports are included.
Operational Rules & Constraints
-
Optimizer Integration:
- Import
fromFusedbun
.optim - In
, instantiateinit_optimizer_state
instead ofFusedbun
.torch.optim.AdamW - Map the following hyperparameters from the input
object to thehyperparameters
constructor:Fusedbun
:lrhyperparameters.learning_rate
:beta_decayhyperparameters.beta_decay
:Lambdahyperparameters.Lambda
:momentum_betahyperparameters.momentum_beta
- Set
andcentralize=True
as defaults.use_rms=True
- Import
-
Scheduler Configuration:
- The
object does not have ahyperparameters
attribute.warmup_factor - In the
function, do not usepytorch_cosine_warmup
.hyperparameters.warmup_factor - Calculate
using a fixed fraction ofwarmup_steps
(e.g.,step_hint
) or remove the warmup logic if specified.warmup_steps = int(0.1 * step_hint) - Ensure
is an integer to preventwarmup_steps
.TypeError: unsupported operand type(s) for -: 'int' and 'tuple'
- The
-
Code Structure:
- Maintain the existing structure of
,update_params
, andget_batch_size
.data_selection - Ensure
is imported fromUSE_PYTORCH_DDP
.algorithmic_efficiency.pytorch_utils
- Maintain the existing structure of
Anti-Patterns
- Do not attempt to access
.hyperparameters.warmup_factor - Do not multiply the
object directly (e.g.,hyperparameters
is invalid).hyperparameters * step_hint
Triggers
- integrate Fusedbun optimizer
- replace AdamW with Fusedbun
- fix warmup_factor error in submission
- algorithmic efficiency submission Fusedbun