Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

MPI tests using > 15 GB memory #626

Closed
quantumsteve opened this issue Sep 7, 2023 · 4 comments
Closed

MPI tests using > 15 GB memory #626

quantumsteve opened this issue Sep 7, 2023 · 4 comments

Comments

@quantumsteve
Copy link
Collaborator

quantumsteve commented Sep 7, 2023

Describe the bug
A clear and concise description of what the bug is.

To Reproduce
Steps to reproduce the behavior:

  1. git commit hash being built
  2. cmake command
  3. full program/test invocation command
  4. additional steps

asgard-unit-mpi-gxx and asgard-unit-mpi-gxx-scalapack currently run out of memory on the CI machines. Iincreased the memory size from 15GB to 144GB as a (temporarily?) workaround, suspect there is something over allocating memory or a test that should be scaled down or moved to a different label that runs less often.

Expected behavior
A clear and concise description of what you expected to happen.

MPI tests pass on a container with 15GB RAM.

System:

  • system name [e.g. fusiont5, summit]
  • modules loaded [e.g. output of module list]
  • other systems where this is reproducible [e.g. "my laptop", "none"]

Additional context
Add any other context about the problem here.

Reproduced locally with docker run -m 15000m -it cpu /bin/bash

@quantumsteve quantumsteve changed the title MPI tests using > 16 GB memory MPI tests using > 15 GB memory Sep 7, 2023
@quantumsteve
Copy link
Collaborator Author

Found while troubleshooting #624.

@quantumsteve
Copy link
Collaborator Author

failing tests are both continuity_6

SECTION("even but not square, large")

SECTION("continuity6, explicit/non-uniform level, degree 4, sparse grid")

@mkstoyanov
Copy link
Collaborator

Did a few tests, I think that 32GB will be enough, but it is still too much for CI.

For some reason, when we create a 6D problem, we are unreasonable amount of memory for something. I wonder if this is an issue with the hash-map taking too much space. If I'm right, this will be an issue across the board, it's just that the problem needs to run 4 copies (since the test uses 4 mpi ranks) and all of our workstations have lots more than that.

@mkstoyanov
Copy link
Collaborator

Tested again on develop, it goes nowhere near 16GB. I suspect the problem was resolved in #743

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants