Skip to content
index.rst 2.12 KiB
Newer Older
Michael Krause's avatar
Michael Krause committed
Introduction
============

This manual describes what the TARDIS is and provides some basic examples and
useful code snippets for various software packages that can be used on LIP's
cluster environment. We are trying to update the documentation every time we
successfully test a new use case. It is meant to be a guideline and if you need
help at any point we will be glad to assist you. We appreciate all of your
feedback so please contact us: `hpc-admin@mpib-berlin.mpg.de
<hpc-admin@mpib-berlin.mpg.de>`_
Michael Krause's avatar
Michael Krause committed

Michael Krause's avatar
Michael Krause committed
Specs
-----
Michael Krause's avatar
Michael Krause committed

Michael Krause's avatar
Michael Krause committed
Tardis
   **T**\ardis, **A** **R**\apid **D**\istributed **I**\nformation **S**\ystem
Michael Krause's avatar
Michael Krause committed

.. image:: ../img/tardis_2019.png
Michael Krause's avatar
Michael Krause committed
   :width: 60%

Some technical facts:
    + **832** Intel® Xeon® CPU E5-2670 CPU cores(no HT) inside 48 Dell m6x0 blade servers
Michael Krause's avatar
Michael Krause committed
    + **R**:sub:`max` = 9.9 TFlops, **R**:sub:`peak` = 14.8 TFlops
    + **8.32TB** total amount of memory
    + **32TB** of attached NFS storage for software
    + **747TB** of BeeGFS storage for user homes
    + fully-connected 100GbE
Michael Krause's avatar
Michael Krause committed


Workflows
---------

**Sequential**

The simplest and very typical processing workflow consist of 3 steps.

+ data download from file servers
+ sequential data processing
+ result upload to file servers

.. image:: ../img/WF-seq.svg
   :width: 80%

Obviously this is not very efficient and with data processing on the Tardis we
would want to do something more efficient. Also this approach requires desktop
machines in a shared working environment to be always on. This procudes
excessive noise and heat. It is also not very nice considering other colleagues
might want to use your machine as well. But most importantly it is **slow**.

**Parallel**

With the Tardis you can login from your laptop or workstation with SSH (see:
:doc:`login`) to a single head node called ``tardis``. On that node users can
prepare and test their code and analysis and then submit it to a queue (see:
:doc:`../pbs/torque`). Jobs will then **eventually** be submitted to one of the
Michael Krause's avatar
Michael Krause committed
computing nodes to get a guaranteed set of processing resources. Afterwards
users can collect the results and copy them back to the file servers.

.. image:: ../img/WF-par.svg
   :width: 80%