-
Notifications
You must be signed in to change notification settings - Fork 93
Deadlock when using MPI with Python packages and DFTK #1067
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
For threading that's just the way it is: For small problems threading includes an overhead. Sometimes less threads is faster than more. Regarding your MPI issue, that looks like a precompilation issue, which can cause MPI-based executions to deadlock. @Technici4n can probably comment whether these symptoms match his experience. @Technici4n we should probably make a note about this in the parallelisation docs. |
For precompilation, you can make sure that all packages are precompiled using For multithreading you can try less threads, for example |
Ok, let's see if this helps to avoid the deadlock for @simonganne01. I any case this should be documented better. |
like this?
because now I get this:
|
Yes exactly. Now it seems you are getting into the fun of trying to figure out what might be wrong with MPI. Without additional logging output it will be difficult. 😅 I would recommend adding print statements at different stages in your code to understand what is failing. (Possibly even the imports). |
Hello everyone,
I'm heaving problems parallelizing DFTK. I run the scripts added to this message. When I run multithreaded vs single-threaded the following is the output:
for multithreaded and
for singlethreaded. As you can see the singlethreaded one is faster. Any idea what is wrong?
For MPI I disable the multithreading and run it with mpiexecjl but it is very slow on startup. I get this output and after that it seems to be stuck:
run_DFTK.jl:
Thanks in advance for the help!
POSCAR.txt
The text was updated successfully, but these errors were encountered: