Overall Resources¶
These are the resources in a list form that are currently (4/12/2018) available to lsst developers.
Machines: (includes VM and Baremetal)¶
lsst-dev01 - login node and submit machine. Currently the “jump” host to get into the NCSA lsst infrasturcture environment.
lsstweb - web services machine
- VM server machines - running IBM’s Vsphere software. (@ NCSA 3003, NCSA NPCF, and Chile Base)
virtual machines for numberous development environments
databackbone test beds, container type test beds, demo type machines, monitoring machines
lsst-demo - VM for demos and docker image needs
ATS gateway - VM
lsstdb - mysql infrastucture machine for db support
DAQ - camera simulator systems
L1 test stand - collection of machines for simulating and testing the camera software
ATS archiver - Auxiliary Telecope Archiver system
Nebula - open stack resource (Use at your own risk)
Slurm Batch cluster - 48 compute nodes for processing DRP and other compute needs
Kubernetes (K8s) - 20 compute nodes for containers
Oracle Rac - 6 nodes for the consolidated DB
PDAC - 35 compute nodes for QSERV and SUI systems
Backup - machines dedicated for data movement to and from DR
Xcat and Puppet configuration management machines - 1 compute node
VM head node - management service for all the VM machines
DTN nodes - data transfer nodes, 2 compute nodes
Monitor DB node and other infrastucture machines (influxDB, loghost…)
identity.lsst.org - reset and password machine
network emulator - latency injection environment for simulating the roundtrip to Chile and back in networks.
Bro cluster - networking scanning software/hardware (Chile Base)
- NCSA provided shared resources :
-Jira, LDAP (ncsa3003, ncsanpcf, Chile base, chile summit), kerberos, IDM, Qualys scans for security, Nearline tape, and Firewall systems, Bro clusters (network scanning security)
Filesystems - in GPFS (4.9PB of storage)¶
/datasets
- Long term storage of project-approved shared data. Contains immuteable data. This is under a disaster recovery policy that every 30 days it is stored and written to nearline tape./home
- Storage of individual-user data. This data is backed up on a daily basis and ncsa retains 30 days of those backups in a snapshot. It does have quotas on this file system for 1TB for each “directory”, and a 1 million INODE quota./software
- Central location for maintenance of project-shared software installations that require access from multiple resources. (ie batch, Nebula)./sui
- Shared storage for ephemeral data for the purpose of supporting SUI/T in the PDAC enclave. This file system has no backups or purging./scratch
- Ephemeral big-data storage for use in computation and other project-related activities. This is not backed up. This file system is purged. Every 30 days a purge policy deleteing files over 180 days./project
- Long term big-data storage for use in computation and other project-related activities. This is backed up with 7 days of snapshots. This file system is not subject to purge.