Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Investigate need for JULIA_CUDA_MEMORY_POOL=none #843

Open
vchuravy opened this issue Jun 19, 2024 · 1 comment
Open

Investigate need for JULIA_CUDA_MEMORY_POOL=none #843

vchuravy opened this issue Jun 19, 2024 · 1 comment

Comments

@vchuravy
Copy link
Member

One datapoint is that locally on OpenMPI 5, the test ran fine on one GPU.

There was a discussion elsewhere (maybe @maleadt remembers) if that flag is still needed or what MPI versions can now handle the new memory interface.

@luraess
Copy link
Contributor

luraess commented Jun 20, 2024

  • From Slack HPC channel (21 days ago)

Tim Besard:
Could somebody who understands CUDA + OpenMPI re-evaluate #537? IIUC, the fact that UCX now supports the CUDA stream-ordered allocator (https://github.com/openucx/ucx/blob/04897a079ac88713842f7209c5e82430d095444e/NEWS#L63) means that this workaround shouldn't be suggested anymore.

The reason being that it is pretty costly, performance wise, and I see it set all the time in HPC user's environments (presumably provided by the system config)

One could (and should) test but isn't UCX one amongst other PML and thus there may be no guarantee that it will just work on clusters not relying on UCX but e.g. libfabric?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants