Request an HPC login here
Please put “Accessing Great Lakes for research projects” in the request.
Log into your account in the web portal
You must connect to U-M VPN to access Great Lakes if you’re not connected to MWireless.
There are three GPU-available partitions.
spgpu
is usually the one with the least wait time, which has 224 NVIDIA A40 GPUs (48GB VRAM). Onspgpu
, you can request 4 CPUs and 48 GB RAM without extra cost for each GPU requested on thespgpu
partition, so the following is the most cost-effective option.
For each GPU requested, Great Lakes provides some CPU cores and RAM memory without additional charge. Please use the following configurations for each partition to maximize the freely-provided CPUs and memory (as of Winter 2025; see the current configurations):
Partition | CPU cores | Memory (RAM) | GPU | GPU speed | GPU memory | #GPUs available |
---|---|---|---|---|---|---|
spgpu | 4 | 48 GB | A40 | faster | 48 GB | 224 |
gpu_mig40 | 8 | 124 GB | A100 | fastest | 40 GB | 16 |
gpu | 20 | 90 GB | V100 | fast | 16 GB | 52 |
standard* | 1 | 7 GB | - | - | - | - |
Here is the cost for each partition (as of Winter 2025; see the current rates):
Partition | Hourly cost | CPU hours equivalent |
---|---|---|
spgpu | 0.11 | 7.33 |
gpu_mig40 | 0.16 | 10.66 |
gpu | 0.16 | 10.66 |
standard* | 0.015 | 1 |
/home/{UNIQNAME}
/scratch/pat463f25_class_root/pat463f25_class/{UNIQNAME}
Select the application (here we take JupyterLab for example)
Specify the configuration of the instance (see below for suggestions)
Connect to the launched instance
This will create a persistent compute node that won’t end automatically. Please make sure to kill the job after your workload is finished so that the resources are released.
Navigate to “Job Composer” on the web portal for Great Lakes or Lighthouse
Click “New Job > From Default Template”
Click “Open Editor”
Replace everything in the editor with the following and click “Save”
This defaults to requesting 1 GPU with 8 CPUs and 62 GB RAM on Lighthouse. See below for configuration suggestions.
#!/bin/bash
#SBATCH --account=aimusic_project
#SBATCH --partition=aimusic_project
#SBATCH --cpus-per-task=8
#SBATCH --mem=62G
#SBATCH --gres=gpu:1
echo $SLURMD_NODENAME
tail -f /dev/null
Click to show an example configuration for GPU workloads on Great Lakes
#!/bin/bash #SBATCH --account=hwdong0 #SBATCH --partition=spgpu #SBATCH --cpus-per-task=4 #SBATCH --mem=48G #SBATCH --gres=gpu:1 echo $SLURMD_NODENAME tail -f /dev/null
In the “Jobs” tab, select the created job and click “Submit”
At the right panel, click the “slurm-{ID}.out” file
Copy the node name
To access the compute node, you will need to first SSH into the login node
ssh {UNIQNAME}@lighthouse.arc-ts.umich.edu
And then you can SSH into the compute node
ssh {NODENAME}.arc-ts.umich.edu
Make sure to delete the job after your workload is finished
You can create a template to save some time from editing the job script every time.
Navigate to the “Templates” tab
Click “New Template”
Enter the template name and click “Save”
Select the created template and click “View Files”
Click the triple dots next to “main_job.sh” and click “Edit”
Replace everything in the editor with the following and click “Save”
This defaults to requesting 1 GPU with 8 CPUs and 62 GB RAM on Lighthouse. See below for configuration suggestions.
#!/bin/bash
#SBATCH --account=aimusic_project
#SBATCH --partition=aimusic_project
#SBATCH --cpus-per-task=8
#SBATCH --mem=62G
#SBATCH --gres=gpu:1
echo $SLURMD_NODENAME
tail -f /dev/null
Click to show an example configuration for GPU workloads on Great Lakes
#!/bin/bash #SBATCH --account=hwdong0 #SBATCH --partition=spgpu #SBATCH --cpus-per-task=4 #SBATCH --mem=48G #SBATCH --gres=gpu:1 echo $SLURMD_NODENAME tail -f /dev/null
Return to the previous page and click “Create New Job”