Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Feature: case insensitive slurm accounts #488

Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
10 changes: 5 additions & 5 deletions group_vars/nibbler_cluster/vars.yml
Original file line number Diff line number Diff line change
Expand Up @@ -30,20 +30,20 @@ additional_etc_hosts:
- calculon-management
- calculon-storage
vcompute_hostnames: "{{ stack_prefix }}-vcompute[01-03]"
vcompute_sockets: 4
vcompute_sockets: 8
vcompute_cores_per_socket: 1
vcompute_real_memory: 3790
vcompute_real_memory: 15884
vcompute_max_cpus_per_node: "{{ vcompute_sockets * vcompute_cores_per_socket - 2 }}"
vcompute_max_mem_per_node: "{{ vcompute_real_memory - vcompute_sockets * vcompute_cores_per_socket * 512 }}"
vcompute_local_disk: 270000
vcompute_local_disk: 975
vcompute_features: 'tmp02'
vcompute_ethernet_interfaces:
- 'eth0'
- 'eth1'
ui_hostnames: "{{ slurm_cluster_name }}"
ui_sockets: 4
ui_sockets: 8
ui_cores_per_socket: 1
ui_real_memory: 3790
ui_real_memory: 15884
ui_local_disk: 0
ui_features: 'prm02,tmp02'
ui_ethernet_interfaces:
Expand Down
2 changes: 1 addition & 1 deletion roles/slurm_management/files/slurm.taskprolog
Original file line number Diff line number Diff line change
Expand Up @@ -13,7 +13,7 @@ set -u

#
# Make sure we have a tmp dir in /local on compute nodes.
# When this failed the job should not continue as SLURM will default to /tmp,
# When this fails, the job should not continue as SLURM will default to /tmp,
# which is not suitable for heavy random IO nor large data sets.
# Hammering /tmp may effectively result in the node going down.
# When the prolog fails the node will be set to state=DRAIN instead.
Expand Down
4 changes: 2 additions & 2 deletions roles/slurm_management/templates/slurm.conf
Original file line number Diff line number Diff line change
Expand Up @@ -131,8 +131,8 @@ HealthCheckInterval=300
# Partitions
#
EnforcePartLimits=YES
PartitionName=DEFAULT State=UP DefMemPerCPU=1024 MaxNodes={% if slurm_allow_jobs_to_span_nodes is defined and true %}{{ groups['compute_vm']|list|length }}{% else %}1{% endif %} MaxTime=7-00:00:01
PartitionName=regular Default=YES MaxNodes={% if slurm_allow_jobs_to_span_nodes is defined and true %}{{ groups['compute_vm']|list|length }}{% else %}1{% endif %} Nodes={{ vcompute_hostnames }} MaxCPUsPerNode={{ vcompute_max_cpus_per_node }} MaxMemPerNode={{ vcompute_max_mem_per_node }} TRESBillingWeights="CPU=1.0,Mem=0.125G" DenyQos=ds-short,ds-medium,ds-long
PartitionName=DEFAULT State=UP DefMemPerCPU=1024 MaxNodes={% if slurm_allow_jobs_to_span_nodes is defined and slurm_allow_jobs_to_span_nodes is true %}{{ groups['compute_vm']|list|length }}{% else %}1{% endif %} MaxTime=7-00:00:01
PartitionName=regular Default=YES MaxNodes={% if slurm_allow_jobs_to_span_nodes is defined and slurm_allow_jobs_to_span_nodes is true %}{{ groups['compute_vm']|list|length }}{% else %}1{% endif %} Nodes={{ vcompute_hostnames }} MaxCPUsPerNode={{ vcompute_max_cpus_per_node }} MaxMemPerNode={{ vcompute_max_mem_per_node }} TRESBillingWeights="CPU=1.0,Mem=0.125G" DenyQos=ds-short,ds-medium,ds-long
PartitionName=ds Default=No MaxNodes=1 Nodes={{ ui_hostnames }} MaxCPUsPerNode=1 MaxMemPerNode=1024 TRESBillingWeights="CPU=1.0,Mem=1.0G" AllowQos=ds-short,ds-medium,ds-long
#
# COMPUTE NODES
Expand Down
1 change: 1 addition & 0 deletions roles/slurm_management/templates/slurmdbd.conf
Original file line number Diff line number Diff line change
Expand Up @@ -25,3 +25,4 @@ StoragePass={{ slurm_storage_pass }}
StorageType=accounting_storage/mysql
StorageUser={{ slurm_storage_user }}
StorageLoc={{ slurm_database_name }}
Parameters=PreserveCaseUser