Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Does the project support multi-gpu training? #216

Open
Snimm opened this issue Oct 3, 2024 · 1 comment
Open

Does the project support multi-gpu training? #216

Snimm opened this issue Oct 3, 2024 · 1 comment

Comments

@Snimm
Copy link

Snimm commented Oct 3, 2024

Does the project support multi-gpu training?
If yes, how? By default, it only uses one GPU. I am unable to find any parameter that can be used for this purpose.

@jonathan-laurent
Copy link
Owner

Although this is not very well tested, AlphaZero.jl will attempt to use Distributed to parallelise some parts of data generation across all available Julia processes. Each process can then be configured to use its own GPU. Evaluation stages are not parallelised this way though and may become a bottleneck if two many processes are used.

Once again, this is not very well tested and documented so may require some digging in. Don't hesitate to report about your experience and contribute some documentation if you manage to make it work for you.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants