Tinaroo

About Tinaroo

Tinaroo is a traditional high performance computing cluster. It is an SGI rackable HPC system for broad research use across UQ.

Tinaroo features an 80 per cent increase in number of processor cores, four-fold increase in average memory and eight-fold peak performance increase over Barrine, the UQ HPC system it is replacing.

Tinaroo has about 6,000 processor cores, and some 30 TBs of main memory. That means it can solve very large problems by splitting it up across the processors and memory modules.

Tinaroo will be focused on tightly coupled parallel jobs to undertake tasks requiring a very large number of computer cores to be applied to a given problem at the same time. It will augment FlashLite, a data-intensive supercomputer and Euramoo, a cloud-hosted virtual cluster. Together these systems will span a wide range of computational templates.

Tinaroo is a UQ HPC resource that has been co-located with QRIScloud and uses QRIScloud's identity management system to ensure interoperability with FlashLite, Euramoo and storage infrastructure. Tinaroo is for UQ staff and students and their research collaborators.

Tinaroo is connected to a file system that provides nearly 0.5 PBs of storage, and has access to QRIScloud and Research Data Services (RDS) data collections. This means researchers can process data they have previously archived using Tinaroo, FlashLite, RDS and/or QRIScloud. The seamless way data can be moved between platforms is very powerful.

Tinaroo, installed in Springfield’s Polaris Data Centre, has access to UQ’s GPFS parallel storage environment, which has been upgraded with an additional capacity of about half-a-petabyte to provide working storage capability.

Getting a Tinaroo account

Access to Tinaroo is through QRIScloud.

To open a Tinaroo account, UQ staff and students should do the following:

Register for a QRIScloud account if they do not already have one (register at QRIScloud).

Click on "Account" and login using their AAF credentials.

All users associated with a university, CSIRO and most other research institutions in Australia will be able to use their organisational ID (login name) and password as their AAF credential without any further process. If a user has no AAF credential, one can be created via QRIScloud's Request an institutional account link.

Tinaroo User Guide

Tinaroo support

Tinaroo Enhanced Access Program

RCC launched a new program in October 2016 to provide enhanced access to HPC Tinaroo for UQ research groups and individuals conducting high impact research.

The benefits of the new mechanism include:

increased queue limits for single node jobs

more flexibility with job array limits

increased maximum allowed wall-times

the ability to submit jobs to Tinaroo from Nimrod, a scientific workflow system

exclusive use of nodes outside of the batch system

virtualisation that permits using the guest operating system of your choice, and more.

See the Tinaroo Enhanced Access Guide for further information about benefits. The mechanism is applicable to the workload, storage and batch system configuration "dimensions".

The application process requires a brief strategic and technical justification for using Tinaroo Enhanced Access. To apply, send an email to rcc-support@uq.edu.au with "Tinaroo Enhanced Access Request" as the subject. Please check the Tinaroo Enhanced Access Guide as to what information is required in your email submission.

Tinaroo technical specifications

There are 61 C2112-4GP3-R-F 2RU (rack-unit) chassis with four server nodes per chassis.

Each of the 244 compute nodes is comprised of:

2 x Intel Xeon E5-2680V3 2.5Ghz, 12 core processors

8 x 16 GB DIMMs (128 GB) [max. 120GB should requested per node]

1 x 1 TB 7.2K SATA drive.

The 244 compute nodes have in aggregate:

488 CPU sockets

5856 CPU cores

31,232 GB of RAM

244 TB of local disk space

Licensed as GPFS storage clients

234.2 TFLOP.

The nodes are connected to each other via a FDR infiniband network arranged in a four-dimensional enhanced hypercube topology. Although the network topology favours jobs running across 18 or fewer nodes (i.e. less than 433 cores), the networking penalty associated with running much larger jobs is not too high. The batch system will be tuned to take full advantage of the network topology.