File tree 6 files changed +0
-157
lines changed
aws-parallelcluster-config/files/default/cloudwatch_agent
aws-parallelcluster-slurm
files/default/head_node_slurm
6 files changed +0
-157
lines changed Original file line number Diff line number Diff line change 552
552
default [ 'cluster' ] [ 'raid_vol_ids' ] = ''
553
553
default [ 'cluster' ] [ 'dns_domain' ] = nil
554
554
default [ 'cluster' ] [ 'use_private_hostname' ] = 'false'
555
- default [ 'cluster' ] [ 'add_node_hostnames_in_hosts_file' ] = node [ 'cluster' ] [ 'use_private_hostname' ]
556
555
default [ 'cluster' ] [ 'skip_install_recipes' ] = 'yes'
557
556
default [ 'cluster' ] [ 'enable_nss_slurm' ] = node [ 'cluster' ] [ 'directory_service' ] [ 'enabled' ]
558
557
default [ 'cluster' ] [ 'realmemory_to_ec2memory_ratio' ] = 0.95
Original file line number Diff line number Diff line change 540
540
],
541
541
"feature_conditions" : []
542
542
},
543
- {
544
- "timestamp_format_key" : " default" ,
545
- "file_path" : " /var/log/parallelcluster/slurm_prolog_epilog.log" ,
546
- "log_stream_name" : " slurm_prolog_epilog" ,
547
- "schedulers" : [
548
- " slurm"
549
- ],
550
- "platforms" : [
551
- " centos" ,
552
- " ubuntu" ,
553
- " amazon"
554
- ],
555
- "node_roles" : [
556
- " ComputeFleet"
557
- ],
558
- "feature_conditions" : [
559
- {
560
- "dna_key" : " use_private_hostname" ,
561
- "satisfying_values" : [" true" ]
562
- }
563
- ]
564
- },
565
543
{
566
544
"timestamp_format_key" : " default" ,
567
545
"file_path" : " /var/log/parallelcluster/clusterstatusmgtd" ,
Load Diff This file was deleted.
Load Diff This file was deleted.
Original file line number Diff line number Diff line change 192
192
action :create
193
193
end
194
194
195
- if node [ 'cluster' ] [ 'add_node_hostnames_in_hosts_file' ] == "true"
196
- directory "#{ node [ 'cluster' ] [ 'slurm' ] [ 'install_dir' ] } /etc/pcluster/prolog.d" do
197
- user 'root'
198
- group 'root'
199
- mode '0755'
200
- end
201
-
202
- cookbook_file "#{ node [ 'cluster' ] [ 'slurm' ] [ 'install_dir' ] } /etc/pcluster/prolog.d/01-pcluster-prolog" do
203
- source 'head_node_slurm/prolog'
204
- owner node [ 'cluster' ] [ 'slurm' ] [ 'user' ]
205
- group node [ 'cluster' ] [ 'slurm' ] [ 'group' ]
206
- mode '0744'
207
- end
208
-
209
- directory "#{ node [ 'cluster' ] [ 'slurm' ] [ 'install_dir' ] } /etc/pcluster/epilog.d" do
210
- user 'root'
211
- group 'root'
212
- mode '0755'
213
- end
214
-
215
- cookbook_file "#{ node [ 'cluster' ] [ 'slurm' ] [ 'install_dir' ] } /etc/pcluster/epilog.d/01-pcluster-epilog" do
216
- source 'head_node_slurm/epilog'
217
- owner node [ 'cluster' ] [ 'slurm' ] [ 'user' ]
218
- group node [ 'cluster' ] [ 'slurm' ] [ 'group' ]
219
- mode '0744'
220
- end
221
- end
222
-
223
195
service "slurmctld" do
224
196
supports restart : false
225
197
action %i( enable start )
Original file line number Diff line number Diff line change @@ -40,19 +40,6 @@ SuspendTimeout=120
40
40
PrivateData=cloud
41
41
ResumeRate=0
42
42
SuspendRate=0
43
- <% if node["cluster"]["add_node_hostnames_in_hosts_file"] == 'true' -%>
44
- #
45
- # PROLOG AND EPILOG
46
- # prolog is executed to add nodes info to /etc/hosts on compute nodes when each job is allocated
47
- # epilog is executed to clean contents written by prolog
48
- # PrologFlags specifies the prolog is executed at job allocation and prologs and epilogs are of different jobs are executed serially
49
- # SchedulerParameters allows jobs to be requeued to other nodes if prolog error exits.
50
- # Note the error exit of prolog drains a node, because the error of prolog is considered as a node error.
51
- Epilog=<%= node['cluster']['slurm']['install_dir'] %> /etc/pcluster/epilog.d/*
52
- Prolog=<%= node['cluster']['slurm']['install_dir'] %> /etc/pcluster/prolog.d/*
53
- PrologFlags=alloc,serial
54
- SchedulerParameters=nohold_on_prolog_fail
55
- <% end -%>
56
43
#
57
44
# TIMERS
58
45
SlurmctldTimeout=300
You can’t perform that action at this time.
0 commit comments