You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In the gsm8k script (link), the distilled dataset is generated using fp16 precision, while the model is trained on this dataset using bf16.
Shouldn't the precision format be consistent throughout the process?
Generate distilled dataset using fp16: [Line 35](https://github.com/sail-sg/sdft/blob/bfb6c255fccdce7459235c20f19a3b9817a9cd5d/scripts/gsm8k/sdft.sh#L35)
Train on distilled dataset using bf16: [Line 61](https://github.com/sail-sg/sdft/blob/bfb6c255fccdce7459235c20f19a3b9817a9cd5d/scripts/gsm8k/sdft.sh#L61)
The text was updated successfully, but these errors were encountered:
Thanks for your interest! In an initial experiment, training with fp16 resulted in instabilities. Consequently, we adopted bf16 for training while continuing to use fp16 for inference. This approach has not led to any significant issues to date.
In the gsm8k script (link), the distilled dataset is generated using fp16 precision, while the model is trained on this dataset using bf16.
Shouldn't the precision format be consistent throughout the process?
The text was updated successfully, but these errors were encountered: