The following benchmarks are selected to be used on the first day of the competition.
HPC Challenge (HPCC) will be used to score the benchmark portion of the competition. A team may execute HPCC as many times as desired during the setup and benchmarking phase, but the HPCC run submitted for scoring will define the hardware baseline for the rest of the competition. In other words, after submitting this benchmark, the same system configuration should be used for the rest of the competition.
The rules described in the Rules section of HPCC web page on code modification does apply.
High Performance LINPACK (HPL)
The teams will compete on High Performance LINPACK (HPL) benchmark for the ‘High LINPACK’ award for the team submitting the highest HPL score. Additional, independent HPL runs (outside the submitted HPCC run) may be considered for the “Highest LINPACK” award if they are performed with exactly the same hardware powered on as used for HPCC run submitted for scoring. While eligible for the Highest LINPACK award, independent HPL runs will NOT count toward the team’s overall score. The HPL run must be submitted on the first day of the competition.
The teams may use any HPL binary.
• The teams need to declare which binary they going to run (by June 5) and provide the binary info + NVIDIA contact (or anyone else) that provided them the binary.
• Due to the Open MPI issue #3003, we advise all student teams to avoid using Open MPI versions between 1.10.3 to 1.10.6 due to the timer bug. This bug can potentially cause HPL to show the calculated results better than the theoretical peak.
HPCG stands for High Performance Conjugate Gradient. It is a self-contained benchmark that generates and solves a synthetic 3D sparse linear system using a local symmetric Gauss-Seidel preconditioned conjugate gradient method. HPCG is a software package that performs a fixed number of symmetric Gauss-Seidel preconditioned conjugate gradient iterations using double precision (64 bit) floating point values. Integer arrays have global and local scope (global indices are unique across the entire distributed memory system, local indices are unique within a memory image). Reference implementation is written in C++ with MPI and OpenMP support. HPCG is being used on the first day of the competition. 30 minutes is the minimum time needed for the official run.
The teams may use any HPCG binary.
Notes: The teams need to declare which binary they going to run (by June 10) and provide the binary info + NVIDIA contact (or anyone else) that provided them the binary.
Cosmological simulation framework “ChaNGa” is a collaborative project with Prof. Thomas Quinn (University of Washington: N-Body Shop) supported by the NSF. ChaNGa (Charm N-body GrAvity solver) is a code to perform collisionless N-body simulations Cosmological simulation. It can perform cosmological simulations with periodic boundary conditions in comoving coordinates or simulations of isolated stellar systems. It also can include hydrodynamics using the Smooth Particle Hydrodynamics (SPH) technique. It uses a Barnes-Hut tree to calculate gravity, with hexadecapole expansion of nodes and Ewald summation for periodic forces. Timestepping is done with a leapfrog integrator with individual timesteps for each particle. ChaNGa runs over Charm++ framework.
In task we will do a a cosmological simulation of a galaxy cluster with an unprecedented resolution.
Elmer/Ice is an Open Source Finite Element Software for Ice Sheet, Glaciers and Ice Flow Modelling. Elmer/Ice is an add-on package to Elmer, which is a multi-physics FEM suite mainly developed by CSC-IT Center for Science Ltd., Espoo, Finland. Initially started by CSC, IGE and ILTS, currently multiple institutions and individuals contribute to the development of Elmer/Ice.
The assignment will be lead by Senior Application Scientist, Professor Thomas Zwinger from HPC CSC in Finland.
HPC Secret Application
A Secret application will be announced on the day of the competition.
Teams, get ready for your interview, meet your judges!
Here are some topics to think about.
1. Each interview is about 5-10 min, mostly on Tuesday (afternoon) or Wednesday (Morning). Focus your answers.
2. Please introduce yourself and the team to the judge.
3. Try that more than one team member will answer the judge. You will need to show team-work.
4. Make sure that the team understands the applications and benchmarks, you will be asked to demonstrate your knowledge.
5. Get to know your HW, Network used and GPUs. What are your considerations for using this cluster architecture.
6. Tuning options: What did you do to tune the applications? What was the considerations?
7. Power Reductions considerations and tuning: What did the team do to stay under the 3KW power limit.
8. Team work: How does each team member participate? Make sure that it is not a one man’s job … we expect all to be involved.
9. Booth design, decoration and general environment. Decorate your booth for points!
10. Overall impressions, mistakes and comments learned.