Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Improved coordination of CUDA contexts with PyTorch #47

Merged
merged 1 commit into from
Nov 9, 2021

Conversation

peastman
Copy link
Member

@peastman peastman commented Oct 4, 2021

Fixes #42. This requires the changes from openmm/openmm#3258. We might want to wait a while before merging it, since it's not compatible with OpenMM 7.6.

@raimis
Copy link
Contributor

raimis commented Oct 6, 2021

cuda-memcheck reports several problems. I'll investigate them further. Meanwhile, I leave openmm/openmm#3272 to you, @peastman.

@raimis
Copy link
Contributor

raimis commented Oct 6, 2021

peastman#1 might be a solution, but let me check after rebuilding everything.

@raimis
Copy link
Contributor

raimis commented Oct 6, 2021

The theory behind:

  • We have two CUDA contexts and a results there two CUDA streams for each
  • Switching between contexts, doesn't synchronise the streams. So an explicit synchronisation is required.

@raimis
Copy link
Contributor

raimis commented Nov 9, 2021

OpenMM 7.7b is out!

How shall we proceed with this? Can conda-forge build the beta of OpenMM-Torch using the OpenMM beta as a dependency?

@peastman
Copy link
Member Author

peastman commented Nov 9, 2021

Sure. I think it should be possible to build a beta that uses the OpenMM beta, though I'm not sure of the details. It would need to use conda-forge/label/openmm-rc as its top priority channel.

@peastman
Copy link
Member Author

peastman commented Nov 9, 2021

Note that CI is failing because it's building against the released OpenMM. I'm going to merge anyway, since it will be necessary for building a beta.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Simulation explodes after adding a TorchForce that is identical to zero
2 participants