Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to run mochi on 2 machine? #122

Open
david-beckham-315 opened this issue Dec 18, 2024 · 1 comment
Open

How to run mochi on 2 machine? #122

david-beckham-315 opened this issue Dec 18, 2024 · 1 comment
Labels
enhancement New feature or request

Comments

@david-beckham-315
Copy link

Hi
There are 2 A100 machines, each has 8 A100 GPUs, does mochi support running on 2 machines with all 16 GPUs? How to run?

Thanks.

@ajayjain
Copy link
Contributor

Hi @david-beckham-315, our repository supports context parallel inference on 1, 2, 4, 6, 8, 12 or 24 GPUs (factors of the number of model heads = 24).

To run inference on 16 GPUs, you'd have to implement another context parallel inference strategy like Ring-Attention. https://github.com/xdit-project/mochi-xdit has a Ring-Attention implementation.

We'd welcome contributions if you get it to work!

@ajayjain ajayjain added the enhancement New feature or request label Dec 20, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants