Elastic HPC Jobs - EFS

Example HPC jobs with CCQ and EFS

Launch a Sample Job to see how CCQ and CloudyCluster Work

To setup the sample job, first ssh into the login instance and copy the sample jobs to the shared file system.

Copy Sample Jobs from /software/samplejobs/ to /mnt/orangefs/samplejobs/

Edit the sample job you would like to run. Remove the extra comment (#) from the scheduler you are running. In this example the Slurm scheduler is used. Uncomment the sharedFS option you selected. Here we are using EFS as an example.

Edit the sample job and uncomment the EFS and Scheduler lines

Launch the job by running the ccqsub command. Check the status with ccqstat:

ccqsub -js mpi_prime2-2.sh

As shown below, the two t2.medium nodes for the jobs are automatically launching

AWS console CCQ instance listing

You can edit the file, save some other node configurations and launch them as well.

ccqsub -js mpi_prime4-1.sh

You can also run spot instance jobs by adding directives to the job script.

#CC -us yes (use spot)
#CC -sp .15 (the price of the bid)
#CC -it c4.2xlarge (the instance type)
ccqsub spot job

Launch the job and check the status.

ccqsub spot job

When the job is in the:

ccqs job status

The completed status is displayed when the job is done.

ccq job status complete

The Job Status Output is sent to the home directory or the directory the job was launched from.

ccq job status output in homedir

When there are no more jobs for that instance type, the instances are terminated.

ccq instances terminating when there are no more jobs

Now you can pause your HPC Environment if you do not have any more jobs to run. Once everything is paused, you can also stop the control node from the AWS Console and restart it when you are ready to run jobs again.

To assist researchers, a wide variety of common open HPC tools and libraries are pre-installed and configured, ready to run and ready to scale.