Replies: 4 comments
-
So the tricky thing about the slurm manager is you are only supposed to run it a single task inside that slurm allocation. It's easiest if you set up an interactive allocation with 2 nodes, and then ssh into the first node and run the script a single time. Btw, with #794 I'm hoping that this will make things a bit easier for "standard" slurm scripts. If you have some time to try it, I'd love to hear if that fixes it for you (run with |
Beta Was this translation helpful? Give feedback.
-
I installed the SlurmClusterManager, but I encountered the following message:
It seems there might be an issue with importing ClusterManagers.addprocs_slurm_native. I am using PySR for Python, and the error is arising in this context. |
Beta Was this translation helpful? Give feedback.
-
You have to install pysr from that branch, since it’s not yet part of the released version yet |
Beta Was this translation helpful? Give feedback.
-
Sorry it turned out it was my fault. I fixed the PR: #794. However, this won't actually work yet, until these upstream PRs are merged:
In the meantime I would recommend trying to get the existing |
Beta Was this translation helpful? Give feedback.
-
I submitted a job to SLURM with the following output:
The job is running, but I'm not getting any output files, and it keeps showing "No output file yet" for a long time.
I am using PySR with the following configuration:
And the job was submitted with the following SLURM command:
sbatch --qos=short --cpus-per-task=128 --nodes=2 ./script.sh
Questions:
Is there any issue with the PySR configuration that could be causing this problem?
Could the issue be with SLURM's resource allocation or the distribution of nodes?
What should I check to ensure that the output files are generated correctly and that the job is using the requested resources properly?
Beta Was this translation helpful? Give feedback.
All reactions