Overall Resources

These are the resources in a list form that are currently (4/12/2018) available to lsst developers.

  1. Machines on the floor.
  2. Storage filesystems.

Machines: (includes VM and Baremetal)

  • lsst-dev01 - login node and submit machine. Currently the “jump” host to get into the NCSA lsst infrasturcture environment.
  • lsstweb - web services machine
  • VM server machines - running IBM’s Vsphere software. (@ NCSA 3003, NCSA NPCF, and Chile Base)
    • virtual machines for numberous development environments
    • databackbone test beds, container type test beds, demo type machines, monitoring machines
  • lsst-demo - VM for demos
  • ATS gateway - VM
  • lsstdb - mysql infrastucture machine for db support
  • DAQ - camera simulator systems
  • L1 test stand - collection of machines for simulating and testing the camera software
  • ATS archiver - Auxiliary Telecope Archiver system
  • Nebula - open stack resource (Use at your own risk)
  • Slurm Batch cluster - 48 compute nodes for processing DRP and other compute needs
  • Kubernetes (K8s) - 20 compute nodes for containers
  • Oracle Rac - 6 nodes for the consolidated DB
  • PDAC - 35 compute nodes for QSERV and SUI systems
  • Backup - machines dedicated for data movement to and from DR
  • Xcat and Puppet configuration management machines - 1 compute node
  • VM head node - management service for all the VM machines
  • DTN nodes - data transfer nodes, 2 compute nodes
  • Monitor DB node and other infrastucture machines (influxDB, loghost…)
  • identity.lsst.org - reset and password machine
  • network emulator - latency injection environment for simulating the roundtrip to Chile and back in networks.
  • Bro cluster - networking scanning software/hardware (Chile Base)
  • NCSA provided shared resources :
    -Jira, LDAP (ncsa3003, ncsanpcf, Chile base, chile summit), kerberos, IDM, Qualys scans for security, Nearline tape, and Firewall systems, Bro clusters (network scanning security)

Filesystems - in GPFS (4.9PB of storage)

  • /datasets - Long term storage of project-approved shared data. Contains immuteable data. This is under a disaster recovery policy that every 30 days it is stored and written to nearline tape.
  • /home - Storage of individual-user data. This data is backed up on a daily basis and ncsa retains 30 days of those backups in a snapshot. It does have quotas on this file system for 1TB for each “directory”, and a 1 million INODE quota.
  • /software - Central location for maintenance of project-shared software installations that require access from multiple resources. (ie batch, Nebula).
  • /sui - Shared storage for ephemeral data for the purpose of supporting SUI/T in the PDAC enclave. This file system has no backups or purging.
  • /scratch - Ephemeral big-data storage for use in computation and other project-related activities. This is not backed up. This file system is purged. Every 30 days a purge policy deleteing files over 180 days.
  • /project - Long term big-data storage for use in computation and other project-related activities. This is backed up with 7 days of snapshots. This file system is not subject to purge.