From 230f2d2822c74c4a4e832c73890c30858839128f Mon Sep 17 00:00:00 2001 From: Bala Desinghu Date: Thu, 6 May 2021 23:29:42 +0000 Subject: [PATCH] adding readme --- README.md | 19 +++++++++++++++++++ 1 file changed, 19 insertions(+) diff --git a/README.md b/README.md index 8ee9d50..6ca22b6 100644 --- a/README.md +++ b/README.md @@ -1,2 +1,21 @@ # CLASS-HPC-GCP Internet2 CLASS Capstone Project, HPC-GCP Team + +t is convenient to deploy an HPC cluster on the GCP cloud using terraform scripts. The scripts are inside the directory slurm-gcp. You need to edit the scripts - basic.tfvars and main.tf. For a basic deployment, it is sufficient to redefine a few variables such as project name, cluster name, etc. Of course, the variables and scripts can be fine tuned to fit your needs. + +Create a GCP Project (for example, class-capstone) +Go to API services and enable the API for compute engines and deployment manager +Start the cloud shell (If you set up the environment for gcloud shell or sdk on your laptop, you can use your development environment in place of cloud shell) +Clone the gcp repo: git clone https://github.com/SchedMD/slurm-gcp.git +Go to the directory slurm-gcp/tf/examples +Make a copy of the basic example. `cp basic.tfvar.example basic.tfvar` +Edit the basics.tfvar file. Add a line “project = class-capstone” (or any name you like) +Open main.tf, and make sure that the bash variable `source` refers to the correct path +Initialize terraform `terraform init` + Start the HPC cluster. `terraform apply -var-file=basic.tfvars`. + Go to your GCP dashboard and check the compute engines. You should see the controller and the login node up and running. + SSH into the login node and check the slurm status (`sinfo`) or run some test jobs. + +Note that the minimal disk-size is 20 GB to accommodate the size of the VM. If you wan’t more than 20 GB, that’s okay. + +