Resource centre HPCG
From HP-SEE Wiki
(Created page with "== '''Resource centre HPCG''' == The HPCG cluster is located at IICT-BAS. It has 576 computing cores organized in a blade system. The storage and management nodes have 128 cores...") |
(→Libraries and application tools) |
||
Line 40: | Line 40: | ||
== ''Libraries and application tools'' == | == ''Libraries and application tools'' == | ||
- | 1) Software libraries: | + | 1) '''Software libraries:''' |
- | ATLAS | + | ATLAS, LAPACK, Linpack, ScaLAPACK, GotoBLAS, FFTW, LUSTRE, SPRNG, MPI (MVIAPICH1/2, OpenMPI), BLACS, BLAS, Maple, VMD, CUDA, OpenCL, OpenFOAM, octave |
- | + | ||
- | LAPACK | + | |
- | + | ||
- | Linpack | + | |
- | + | ||
- | ScaLAPACK | + | |
- | + | ||
- | GotoBLAS | + | |
- | + | ||
- | FFTW | + | |
- | + | ||
- | LUSTRE | + | |
- | + | ||
- | SPRNG | + | |
- | + | ||
- | MPI (MVIAPICH1/2, OpenMPI) | + | |
- | + | ||
- | BLACS | + | |
- | + | ||
- | BLAS | + | |
- | + | ||
- | Maple | + | |
- | + | ||
- | VMD | + | |
- | + | ||
- | CUDA | + | |
- | + | ||
- | OpenCL | + | |
- | + | ||
- | OpenFOAM | + | |
- | + | ||
- | octave | + | |
2) Development and application software available: | 2) Development and application software available: | ||
- | Charm++ | + | Charm++, CPMD, GAMESS, GROMACS, NAMD, NWChem, Quantum Espresso, mpiBLAST, WRF, CMAQ, SMOKE |
- | + | ||
- | CPMD | + | |
- | + | ||
- | GAMESS | + | |
- | + | ||
- | GROMACS | + | |
- | + | ||
- | NAMD | + | |
- | + | ||
- | NWChem | + | |
- | + | ||
- | Quantum Espresso | + | |
- | + | ||
- | mpiBLAST | + | |
- | + | ||
- | WRF | + | |
- | + | ||
- | CMAQ | + | |
- | + | ||
- | SMOKE | + |
Revision as of 10:30, 15 March 2011
Contents |
Resource centre HPCG
The HPCG cluster is located at IICT-BAS. It has 576 computing cores organized in a blade system. The storage and management nodes have 128 cores. All these servers are interconnected via non-blocking DDR Infiniband interconnect. A smaller cluster with powerful GPU computing cards is also attached to it. The main scientific areas covered by this center are environmental modelling, computational physics, computational chemistry and biomedicine.
Parallel programming models
Parallel programming paradigms supported by HPCG cluster are Message passing, supporting several implementations of MPI: MVIAPICH1/2, OpenMPI, OpenMP, as a shared memory approach, available through GNU Compiler Collection (GCC). The nodes have relatively high amount of RAM (24GB per node). Hybrid approach available through combining the two approaches listed above. GPU computing is available by using CUDA and/or OpenCL
Development software
Several versions of the GCC toolchain are available, in order to have flexibility and resolve portability issues for some software packages. Performance and debugging tools include standard gdb and gprof as well as MPE, mpiP and SCALASCA.
Infrastructural services
User administration, authentication, authorization, security
The main way to use the cluster for HPC work is by standard authentication through username/password or public key authentication. It is also possible to submit jobs using the gLite Grid middleware, provided the user has X.509 certificate and is a member of an appropriate supported Virtual Organization.
Workflow and batch management
HPCG cluster is using a Torque + Maui combination. The main way to manage resource utllizations is from the Maui configuration.
Distributed filesystem and data management
There are two main filesystems (/home and /gscratch), both based on the high performance Lustre filesystem. The latter is used only for temporary, but large files, created during job execution.
Accounting and resource management
A custom solution that gathers accounting data from several Bulgarian clusters has been developed and deployed. This solution provides accurate low-level data for all jobs run at these clusters and may be used not only for aggregate accounting but also for performance monitoring.
Operational monitoring
Monitoring of HPCG cluster is performed through the nagios portal. Some additional tools like pakiti are also available.
Helpdesk and user support
User support is achieved through email lists or through the regional helpdesk.
Libraries and application tools
1) Software libraries:
ATLAS, LAPACK, Linpack, ScaLAPACK, GotoBLAS, FFTW, LUSTRE, SPRNG, MPI (MVIAPICH1/2, OpenMPI), BLACS, BLAS, Maple, VMD, CUDA, OpenCL, OpenFOAM, octave
2) Development and application software available:
Charm++, CPMD, GAMESS, GROMACS, NAMD, NWChem, Quantum Espresso, mpiBLAST, WRF, CMAQ, SMOKE