You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 50 Next »

Quick Start


Contact Us

Request for HAL Access

Please fill out the following form. Make sure to follow the link on the application confirmation page to request an actual system account.

Application Form
https://forms.illinois.edu/sec/6587313

Submit Tech-Support Ticket

Please submit a tech-support ticket to the admin team.

Tech-Support Mailbox
help+isl@ncsa.illinois.edu

Join HAL Slack User Group

Please join HAL slack user group.

HAL Slack User Group
https://join.slack.com/t/halillinoisncsa

Check System Status

Please visit the following website to monitor real-time system status.

HAL Real-time System Monitor
https://hal-monitor.ncsa.illinois.edu:3000/

User Guide

Connect to HAL cluster

There are 2 methods to log on to the HAL system. The first method is to SSH via terminal,

SSH
ssh <username>@hal.ncsa.illinois.edu

and the second method is to visit the HAL OnDemand webpage.

HAL OnDemand
https://hal.ncsa.illinois.edu:8888

Submit Interactive Job to HAL

There are 2 ways to submit interactive jobs to the HAL system. The first one is to use the Slurm Wrapper Suite,

swrun -p gpux1

and the second method is to submit with Slurm directly.

srun --partition=gpux1 --pty --nodes=1 --ntasks-per-node=12 \
  --cores-per-socket=3 --threads-per-core=4 --sockets-per-node=1 \
  --gres=gpu:v100:1 --mem-per-cpu=1500 --time=2:00:00 \
  --wait=0 --export=ALL /bin/bash

Submit Batch Job to HAL

There are 2 ways to submit batch jobs to the HAL system. The first one is to use the Slurm Wrapper Suite,

swbatch run_script.swb

The run_script.swb example

run_script.swb
#!/bin/bash

#SBATCH --job-name="hostname"
#SBATCH --output="hostname.%j.%N.out"
#SBATCH --error="hostname.%j.%N.err" 
#SBATCH --partition=gpux1

srun /bin/hostname # this is our "application"

and the second method is to submit with Slurm directly.

swbatch run_script.sb

The run_script.sb example

#!/bin/bash

#SBATCH --job-name="hostname"
#SBATCH --output="hostname.%j.%N.out"
#SBATCH --error="hostname.%j.%N.err" 
#SBATCH --partition=gpux1 
#SBATCH --nodes=1
#SBATCH --ntasks-per-node=1 
#SBATCH --export=ALL 
#SBATCH -t 00:10:00 

srun /bin/hostname # this is our "application"

Documentation


System Overview

Hardware Information

Login node

Login NodeIBM9006-12P1x
CPUIBMPOWER9 16 Cores2x
NetworkMellanox2 Ports EDR InfiniBand1x

Compute node

Compute NodeIBMAC922 8335-GTH16x
CPUIBMPOWER9 20 Cores2x
GPUNVidiaV100 16GB Memory4x
NetworkMellanox2 Ports EDR InfiniBand1x

Storage Node

Storage NodeIBM9006-22P1x
CPUIBMPOWER9 20 Cores2x
StorageWDNFS1x
NetworkMellanox2 Ports EDR InfiniBand1x

Software Information

ManufacturerSoftware PackageVersion
IBMRedHat Linux7.6
NVidiaCUDA10.1.105
NVidiaPGI Compiler19.4
IBMAdvance Toolchain12.0
IBMXLC/XLF16.1.1
IBMPowerAI1.6.1
SchedMDSlurm19.05.2
OSCOpen OnDemand1.6.20

Job Management with Slurm

Frequently Asked Questions


  • No labels