You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hey Genmo team, thank you so much for the open-source of the fine-tuning code.
Are there any plans for releasing a multi-gpu fine-tuning version?
Was there a specific reason there is only a single-gpu fine-tuning code release? Is it nontrivial to make it multi-gpu training?
Thanks in advance!
The text was updated successfully, but these errors were encountered:
Hi @joeljang, at the moment we're not planning to release a multi-GPU trainer. One of the top community requests has been to reduce the minimum resource requirements for using and fine-tuning Mochi, so we made sure the LoRA tuner could run on a single GPU.
@vedantroy Following up on this -- wouldn't it just be a matter of adding DDP? As forward + backward fit fully on a single GPU, shouldn't of sharding data and averaging gradient updates be sufficient?
Hey Genmo team, thank you so much for the open-source of the fine-tuning code.
Are there any plans for releasing a multi-gpu fine-tuning version?
Was there a specific reason there is only a single-gpu fine-tuning code release? Is it nontrivial to make it multi-gpu training?
Thanks in advance!
The text was updated successfully, but these errors were encountered: