IU's campus is the home to three supercomputers: Big Red II, Karst, and Mason.
If you are a new user looking to join a existing HPC project, do the following:
Get added to the RTProject https://projects.rt.iu.edu/
This should be done by a admin on the project (Ehren is likely this person)
You need to be to the project as a 'user'
Scroll down to 'users' and click '+ user', search for email and click okay
You also need to be connected to the allocations of the project, such as 'Quartz' to use it
Scroll down to 'Allocations' and click the user icon on the far right
Select the checkbox next to the user and click 'add to allocation'
Double check that they have read/write permission, the default seems to be read only.
Get an account on the respective resources (This should probably be done first)
https://access.iu.edu/Accounts/Create - this link will help you set up your accounts
You will need accounts on 'Quartz' and 'Slate-Project' To get these, you need to select one and follow the prompts and then start again for the other one.
FOR Slate-Project - It will ask if you are storing HIPAA protected data - the answer is 'NO'
If there are technical problems, email hpfs-incoming@rt-tickets.uits.iu.edu
There are three things that must be true for a computational project to run on HPC.
A 'RT Project' - stands for a Research Technology Project
These are administered through https://projects.rt.iu.edu/
These are effectively 'accounts' to which compute time and storage space are charged (though at no current $ cost).
What this means in practice - before storage space can be allocated or a jobs can be run, it must be allocated to a project
A user must be added to a project before they can run jobs that charge compute time to the project
Big Red II runs large parallel and data-intensive jobs.
hostname: bigred2.uits.iu.edu
Karst runs serial jobs and small parallel jobs. Users also have the option to purchase dedicated nodes.
hostname: karst.uits.iu.edu
Mason supports life sciences and bioinformatics jobs that require large amounts of memory.
hostname: mason.indiana.edu
Login command:
ssh <username>@<hostname>
Example: ssh joe@mason.indiana.edu
Password: IU CAS password
IU's HPC systems use the Modules environment management package. The Modules system helps you add and remove programs to the session to make them available to run.
Display available modules
module av
Display loaded modules
module list
Search for software
module av gromacs
Load module
module load <name/of/module>
If a load command returns a conflict, you can unload the conflicting module
module unload <name/of/conflicting/module>
For more about Modules, see "On Big Red II, Karst, and Mason at IU, how do I use Modules to manage my software environment?" kb.iu.edu/d/bcwy
The Karst and Mason systems run the Red Hat Enterprise Linux 6.0 (RHEL) operating system. Big Red II runs the Cray Linux Environment (CLE) operating system.
Java and MPI wrapper compilers (Open MPI and MPICH) are also available.
For more information, see "What compilers are available on the IU research systems?" kb.iu.edu/d/abby
Job Management is provided by the TORQUE resource in combination with the Moab job scheduler. Batch scripts will need to be configured for the specific system upon which the script will run.
A TORQUE batch job script for an MPI job no Karst might look like this:
#!/bin/bash
#PBS -l nodes=2:ppn=16,walltime=30:00
#PBS -M jthutt@tatooine.net
#PBS -m abe
#PBS -N JobName
#PBS -j oe
mpirun -np 32 -machinefile $PBS_NODEFILE a.out
For more about job scripts, see "What is Torque, and how do I use it to submit and manage jobs on high-performance computing systems?" kb.iu.edu/d/avmy
EXAMPLE JOB CONTROL COMMANDS
Submit script:
qsub my_job_script.pbs
Query scheduler:
showq -i
Query job status:
qstat -u username
Delete a submitted job:
qdel job_id
For more about scheduling, see "What is Moab?" kb.iu.edu/d/avmu
Note: Big Red II is made up of two parts: the external service nodes and the actual compute nodes where the jobs run. The figure to the right shows that all jobs on Big Red II will first land on an "aprun" service node. Jobs require unique commands to launch jobs on Big Red II compute nodes. In general, use the "aprun" command to launch your jobs, whether they are serial or parallel.
For more information, see "Execution environments on Big Red II at IU: Extreme Scalability Mode (ESM) and Cluster Compatibility Mode (CCM)" kb.iu.edu/d/bdol and "How do I run batch jobs on Big Red II at IU?" kb.iu.edu/d/bdkt
For more information about queues, see "Queue information for IU research and computing systems" kb.iu.edu/d/bdkd
In addition to fire-and-forget jobs, you can interactively run jobs on the compute nodes.
To run interactive jobs, use "qsub -I" along with other command line arguments to describe your job needs,
On Karst, use the interactive queue. For example:
qsub -I -q interactive -l nodes=1:ppn=1,vmem=10gb,walltime=4:00:00
To use X forwarding in an interactive job, use the -X flag. For example, on Karst:
qsub -I -X -l walltime=00:45:00 -l nodes=1:ppn=4 -l vmem=10gb
For more information about X forwarding, see "On my personal computer, how do I use X forwarding to securely run graphical application installed on Big Red II, Karst, or Mason at IU? kb.iu.edu/d/bdnt
When you create an account on one of IU's research computing resources, you are allotted 100 GB of home directory disk storage. The 100 GB quota is shared across all accounts on IU machines.
An example path to you home directory is as follows: /N/u/<username>/BigRed2
For more about home directories, see "At IU, how much disk space is available to me on the research computing systems?" kb.iu.edu/d/avkm
Data Capacitor II File System
Data Capacitor II (DC2) is a large-capacity, high-throughput, high-bandwidth Lustre file system serving all IU campuses. It is mounted on Big Red II, Karst, and Mason.
For more about DC2, see "At IU, where is my scratch or project directory on the Data Capacitor II or DC-WAN file system, and how do I view directory permissions and determine how much space I"m using?" kb.iu.edu/d/axod
You can access your storage space on DC2 at /N/dc2/scratch/<username>. This space is not intended for permanent storage of data and is not backed up.
For more about high-speed file systems, see "The Data Capacitor II and DC-WAN high-speed file systems at Indiana University" kb.iu.edu/d/avvh
USEFUL LUSTRE COMMANDS
Check storage space usage (get a total sum of data stored in your scratch directory)
du -hc /N/dc2/scratch/<username>
Find oldest files (number list files in reverse order by date modified)
$ cd /N/dc2/scratch/<username>
$ find . -type f -exec ls -hltr "{} +;y
For more, see "About Lustre file systems" kb.iu.edu/d/ayfh
The Scholarly Data Archive
The Indiana University Scholarly Data Archive (SDA) provides extensive capacity (42 PB) for storing and accessing research data. Files that need to be kept for a long time but are not frequently updated are best suited for storage here.
Methods available for transferring data to and from the SDA include Kerberos-enables FTP, parallel FTP (pftp_client), Hierarchical Storage Interface (HSI), secure FTP (SFTP), secure copy (SCP), SMB/CIFS/Windows file sharing (SMB), and https (via a web browser).
For more information about the SDA, see "The Scholarly Data Archive (SDA) at Indiana University" kb.iu.edu/d/aiyi
IU Cyberinfrastructure Gateway
https://cybergateway.uits.iu.edu/iugateway
The IU Cyberinfrastructure Gateway provides a centralized web interface where you can:
Monitor you usage of IU research computing and storage resources: jobs/month, core hours/month, queue time/month, storage on RFS and SDA, etc.
Move data from desktop of laptop workstations to the IU SDA and DC2
Find information about available scientific and bioinformatics applications