Slurmprolog and epilog
Starting with AWS ParallelCluster version 3.6.0, the Slurm configuration that's deployed with AWS ParallelCluster
includes the Prolog and Epilog configuration parameters:
# PROLOG AND EPILOG Prolog=/opt/slurm/etc/scripts/prolog.d/* Epilog=/opt/slurm/etc/scripts/epilog.d/* SchedulerParameters=nohold_on_prolog_fail BatchStartTimeout=180
For more information, see the Prolog and Epilog Guide
AWS ParallelCluster includes the following prolog and epilog scripts:
-
90_plcuster_health_check_manager(in thePrologfolder) -
90_pcluster_noop(in theEpilogfolder)
Note
Both the Prolog and Epilog folder must contain at least one file.
You can use your own custom prolog or epilog scripts by adding them to
the corresponding Prolog and Epilog folders.
Warning
Slurm runs every script in the folders, in reverse alphabetical order.
The run time duration of the prolog and epilog scripts impact the time needed to run a job.
Update the BatchStartTimeout configuration setting when running multiple or long running
prolog scripts. The default is 3 minutes.
If you are using custom prolog and epilog scripts, locate the scripts in the respective
Prolog and Epilog folders.
We recommend that you keep the 90_plcuster_health_check_manager script that runs before every custom script.
For more information, see Slurm configuration customization.