-
Notifications
You must be signed in to change notification settings - Fork 57
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Node states are unknown #37
Comments
Yes, I can reproduce in 21.08, but 20.11.8 appears to be working. Something related to networking changed in Slurm between those releases and the Help is appreciated 😄 |
So I actually found a solution; removing redundant NodeHostName statements in slurm.conf seems to work:
Not that I could find anything in the docs that would suggest this solution; I basically reacted to this error in the logs:
I'm currently running the test on my fork (see changeset). I'll submit a pr on success. |
Fixed in #38 |
Hi,
I'm using docker-centos7-slurm to test a workflow manager. It has been a while since updating, but when trying out the most recent version, I notice that only one node (
c1
) is up in the container. I am currently testing this in my fork (see pr #1). Briefly, I parametrized test_job_can_run to pass partition to the--partition
option. Thenormal
partition works as expected, butdebug
fails.If one enters the latest image with
running
sinfo
yieldsSee github action resuts, where I added some print statements to see what was going on (nevermind that the test actually passed; I was simply looking at the erroneous slurm output file). I consistently get the feedback that the required nodes are not available; it would seem node
c1
is the only node available to sbatch.Are you able to reproduce this?
Cheers,
Per
The text was updated successfully, but these errors were encountered: