Rivanna Status

HPC

Afton is the University of Virginia's newest High-Performance Computing system. The Afton supercomputer is comprised of 300 compute node each with 96 compute cores based on the AMD EPYC 9454 architecture for a total of 28,800 cores. The increase in core count is augmented by a significant increase in memory per node compared to Rivanna. Each Afton node boasts a minimum of 750 Gigabytes of memory, with some supporting up to 1.5 Terabytes of RAM memory. The large amount of memory per node allows researchers to efficiently work with the ever-expanding datasets we are seeing across diverse research disciplines. The Afton and Rivanna systems provide access to 55 nodes with NVIDIA general purpose GPU accelerators (RTX2080, RTX3090, A6000, V100, A40, and A100), including an NVIDIA BasePOD. The Afton and Rivanna platforms share storage systems of over 8 Petabytes and a stack of pre-installed software packages available for computational research across many disciplines.

Rivanna is one of the University of Virginia's High-Performance Computing (HPC) systems. After hardware upgrades and expansions in 2019, the Rivanna supercomputer currently has 603 nodes with over 20,476 compute cores. The Rivanna and Afton systems provide access to 55 nodes with NVIDIA general purpose GPU accelerators (RTX2080, RTX3090, A6000, V100, A40, and A100), including an NVIDIA BasePOD. Both platforms share storage systems of over 8 Petabytes and a stack of pre-installed software packages available for computational research across many disciplines.

Facilities Statement - Are you submitting a grant proposal and need standard information about UVA research computing environments? Get it here.

The sections below contain important information for new and existing Rivanna and Afton users. Please read each carefully.

New users are invited to attend one of our free orientation sessions ("Introduction to HPC") held throughout the year.


Get Started

Access / Allocations

Learn how to request an allocation and add collaborators.

Logging In

Log in through a Web browser or a command-line tool.

File Transfer

Moving files between Rivanna/Afton and other systems.

Software

See a listing of available software.

Storage

Options for free short-term and leased long-term storage

Running Jobs in Slurm

Submitting jobs to Rivanna/Afton through the Slurm resource manager

Job Queues

Determine the best queue (or “partition”) for running your jobs.

Usage Policies

Understand the terms and conditions for using Rivanna/Afton.

FAQs

Frequently Asked Questions.


Overview

A high performance computing cluster is typically made up of at least four service layers:

  1. Login nodes - Where you log in, interact with data and code, and submit jobs.
  2. Compute nodes - Where production jobs are run. On Rivanna and Afton these nodes are heterogenous; some have more memory, some have GPU devices, and so forth. Partitions are homogeneous so you can select specialty hardware by your partition request, sometimes along with a resource request (gres).
  3. Storage - Where files are stored, accessible by all nodes in the cluster.
  4. Resource Manager - A software system that accepts job requests, schedules the jobs on a node or set of nodes, then manages their execution.

Click on elements of the image to learn more:

Parts of a High Performance Computing cluster Login Nodes Compute Node Storage Resource Manager

System Details

Hardware Configuration

SystemCores/NodeMemory/NodeSpecialty HardwareGPU memory/DeviceGPU devices/Node# of Nodes
Rivanna40375GB---182
Afton96750GB---294
Rivanna40750GB---39
Afton961464GB---6
Afton128976GBGPU: A10040GB82
Afton1281953GBGPU: A10080GB817
Afton961464GBGPU: A4048GB86
Afton96750GBGPU: A4048GB86
Afton48250GBGPU: A600048GB814
Afton40374GBGPU: RTX2080Ti11GB102
Afton32125GBGPU: RTX309024GB44
Afton36375GBGPU: V10032GB42
Facilities Statement - Are you submitting a grant proposal and need standard information about UVA research computing environments? Get it here.

Job Queues

Rivanna and Afton are managed resources; users must submit jobs to queues controlled by a resource manager, also known as a queueing system. The manager in use on Rivanna and Afton is Slurm. Slurm refers to queues as partitions because they divide the machine into sets of resources. There is no default partition and each job must request a specific partition. Partitions and access policies are subject to change, but the following table shows the current structure. Note that memory may be requested per core or for the overall job. If the total memory required for the job is greater than the number of cores requested multiplied by the maximum memory per core, the job will be charged for the additional cores whether they are used or not. In addition, jobs running on more than one core may still require a request of total memory rather than memory per core, since memory per core is enforced by the system but some multicore software packages (ANSYS, for example) may exceed that for a short time even though they never exceed cores x memory/core.

PartitionPurposeMax time / jobMax nodes / jobMax cores / jobMax cores / nodeDefault memory / coreMax memory / node / jobSU Charge Rate
standardFor jobs on a single compute node7 days140969GB1462GB1
parallelFor large parallel jobs on up to 64 nodes3 days646000968GB750GB1
gpuFor jobs using general purpose graphical processing units3 days432GPU1286GB1953GB3
interactiveFor quick interactive sessions12 hours224/2GPU966GB216GB1

Remarks

  • standard maximum aggregate CPU cores allowed for a single user’s running jobs is 1000.
  • parallel requires at least 2 nodes and 4 CPU cores.
  • Slurm’s default memory unit is in MB. Different units may be specified, e.g. --mem=100G, where 1G = 1024M.
  • The gpu partition is dedicated to jobs that can utilize a general purpose graphics processing unit (GPGPU). In Slurm scripts you must request at least one GPU device through --gres=gpu. Jobs that do not utilize any GPUs are not allowed in this partition.
    The maximum aggregate GPUs allowed for a single user’s running jobs is 32.
    SU charge rate = #cores + 2 x #GPUs, which is at least 1 + 2x1 = 3.
  • interactive maximum aggregate CPU cores (GPUs) is 24 (2) for a single user. The charge rates for cores and GPUs are the same as the other partitions.
  • The NVIDIA DGX BasePOD offers high-performance GPUs that brings new AI and ML functionality to support parallel GPU computing and large deep-learning models.
      

Usage Policies

Research computing resources at the University of Virginia are for use by faculty, staff, and students of the University and their collaborators in academic research projects. Personal use is not permitted. Users must comply with all University policies for access and security to University resources. The HPC system has additional usage policies to ensure that this shared environment is managed fairly to all users. UVA’s Research Computing (RC) group reserves the right to enact policy changes at any time without prior notice.

Login Nodes

Exceeding the limits on the login nodes (frontend) will result in the user’s process(es) being killed. Repeated violations will result in a warning; users who ignore warnings risk losing access privileges.

Scratch Directory

UVA HPC's scratch file system has a limit of 10TB per user. This policy is in place to guarantee the stability and performance of the scratch file system. Scratch is intended as a temporary work directory. It is not backed up and files that have not been accessed for more than 90 days are marked for deletion. Users are encouraged to back up their important data. Home directories and leased storage are not subject to this policy.

Software Licenses

Excessive consumption of licenses for commercial software, either in time or number, if determined by system and/or RC staff to be interfering with other users' fair use of the software, will subject the violator’s processes or jobs to termination without warning. Staff will attempt to issue a warning before terminating processes or jobs but inadequate response from the violator will not be grounds for permitting the processes/jobs to continue.

Inappropriate Usage

Any violation of the University’s security policies, or any behavior that is considered criminal in nature or a legal threat to the University, will result in the immediate termination of access privileges without warning.