-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
32 b #121
base: main
Are you sure you want to change the base?
Conversation
@@ -130,7 +130,7 @@ def build(self, trainer: "Trainer") -> Optional[Callback]: | |||
eval_batch_size = ( | |||
self.eval_batch_size | |||
if self.eval_batch_size is not None | |||
else trainer.rank_microbatch_size * get_world_size(trainer.dp_process_group) | |||
else 2 * trainer.rank_microbatch_size * get_world_size(trainer.dp_process_group) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
nit: you could instead passed an updated evaluator callback in OLMo2-32B.py
:
.with_callback(
"lm_evaluator",
LMEvaluatorCallbackConfig(
eval_batch_size=<whatever you want>,
eval_dataset=NumpyDatasetConfig.from_data_mix(
DataMix.v3_small_ppl_validation,
name=NumpyDatasetType.padded_fsl,
mix_base_dir=root_dir,
sequence_length=dataset_config.effective_sequence_length,
tokenizer=tokenizer_config,
work_dir=get_work_dir(root_dir),
),
eval_interval=1000,
),
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yeah, but I think this is better. I think we can default to 2x the training batch size. It should always work.
return ( | ||
TrainerConfig( | ||
save_folder=common.save_folder, | ||
save_folder=f"gs://ai2-llm/checkpoints/{project_name}/", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Why change this?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
It defaults to something under my name? Not what we want for an official run?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Especially if we swap babysitting responsibilities during the run
# import flash_attn.ops.triton.cross_entropy as flash_attn_ce # type: ignore | ||
|
||
_fused_cross_entropy_loss = triton_ce_loss.cross_entropy_loss | ||
import flash_attn.ops.triton.cross_entropy as flash_attn_ce # type: ignore |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Our in-house triton CE loss was copied directly from the flash-attn repo, so I don't see the point of this.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ok, I took this back out.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Do I want compiling and fused loss at the same time?
""" | ||
d_model = 5120 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
this is a very narrow model then... are you sure about that?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
It's a clone of Qwen 32. The tradeoffs are, narrow d_model, wide FFN, GQA, lots of layers.
src/scripts/train/OLMo2-32B.py
Outdated
fused_loss=True, | ||
compile_loss=False, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I understand the trepidation about the different loss implementations, but the way it was before was the most performant. This way will be slower and have a higher memory footprint.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Can we have some certainty that this will do the right thing? What happens if we take the 13B from a late checkpoint and run it?
enabled=False, | ||
cancel_check_interval=10, | ||
), | ||
).with_callback( |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We should just add this to the common callbacks.
"lm_evaluator": LMEvaluatorCallbackConfig( |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I don't know that we want these for everything. Default should probably be only the new, blessed ones.
@@ -85,10 +94,57 @@ def build_trainer_config(common: CommonComponents) -> TrainerConfig: | |||
WandBCallback( | |||
name=common.run_name, | |||
entity="ai2-llm", | |||
project="OLMo-core-26B", | |||
project=project_name, | |||
enabled=False, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Intentionally disabled still? Just checking
No description provided.