Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

Table of Contents

We offer a shared condominium cluster Sol two High Performance Computing clusters, and a Ceph-based storage cluster for the Lehigh Research Community.

  • Sol: a 8793-node High Performance Computing ClusterComputing Cluster. Sol is a condominium cluster and can be expanded by interested research groups by investing in the Condo Program.
  • Hawk: a 34-node High Performance Computing Cluster funded by the National Science Foundation's Campus Cyberinfrastructure Award 2019035
  • Ceph: a 768TB combined 2019TB storage cluster based on the Ceph distributed storage system. 

The computing and data storage resources are available for use in the classroom for registrar scheduled courses. See Teaching Uses below.

Download Flyers HPC and Services

Info
titleHPC Accounts

All users of Lehigh's research computing systems must obtain an account for their exclusive use with their

...

personal Lehigh University

...

username.


Info
titleAcknowledgement of Lehigh Research Computing Resources

In publications, reports, and presentations that utilize Research Computing Services, please acknowledge Lehigh University  with the following statement:

"Portions of this research were conducted with research computing resources provided by Lehigh University"

Also, providing the Research Computing group with a list of publications, students graduated, their thesis/dissertation title, and grants received based on work done on Research Computing resources is strongly recommended.

...

An HPC account provides access to Sol , a 87-node condo cluster for researchand Hawk. A faculty member may obtain access to Sol and Hawk by purchasing submitting a minimum allocation of 50,000 core-hours or service units (SU) for $500. Additional allocations can be purchased in increments of 10,000 SUs for $100 each. An allocation cycle begins on Oct. 1 and ends on Sep. 30 of the next year and will not roll over to the next allocation cycle. At launch, there are 1,400,000 SUs available per allocation cycle. Faculty members who require more computing time or want a guaranteed share of total available computing time annually should consider a Condo Investment.The faculty member or PI can request additional accounts sharing the same allocation for an annual charge of $50 (each). Each service level 2 user has a home directory quota of 150GB.proposal for allocation to the Research Computing Steering Committee (RCSC). There are several types of allocation depending on your needs as described in the Accounts & Allocation page.  


Account Request: To obtain an HPC account or allocation or both on Sol, the Faculty sponsoring the user account should contact submit an allocation request using on the forms available at Accounts & Allocation page. If a Faculty already has an active allocation, then please email the Manager of Research Computing for more informationto add additional users to your allocation.

Ceph Storage

In Fall 2018, Research Computing deployed a refreshed 768TB Ceph Storage solution. Faculty, Departments, Centers and Colleges can purchase a Ceph project in units of 1TB.   In Fall 2020, Ceph was augmented by an additional 796TB from Hawk and 455TB by LTS for a total of 2PB storage cluster.

  • Data is replicated on across three disks on three nodes, secured against simultaneous failure of two full nodes in the EWFM cluster.
  • Ceph software performs self-healing (maintaining three replicas) if one or two replicas are lost due to disk or node failure.
  • Ceph software performs daily and weekly data scrubbing to ensure replicas remain consistent to avoid bit rot. 
  • Data which is deleted is NOT recoverable.
  • Data is NOT protected against catastrophic cluster failures or loss of the EWFM datacenter.

Research groups can opt for backup of data to a secondary Ceph cluster in which case

  • Data is stored on two distinct clusters in two locations. The primary cluster is located in the EWFM datacenter, the backup in the Packard datacenter.
  • Data is replicated on across three disks on three nodes in each cluster, secured against simultaneous failure of two full nodes in either cluster or simultaneous failure of five nodes in both clusters.
  • Ceph software performs self-healing (maintaining three replicas) if one or two replicas are lost due to disk or node failure.
  • Ceph software performs daily and weekly data scrubbing to ensure replicas remain consistent to avoid bit rot.
  • Data is snapshotted and stored weekly.
  • Data is protected in the event of catastrophic failure of the primary cluster or loss of the EWFM datacenter, so long as the Packard site remains operational.

Ceph Charges

...

Ceph Charges

  • There are no charges if requesting a storage allocation from the RCSC. Storage allocations that are not renewed or approved annually will be deleted and no backups will be kept.
  • Ceph projects can be purchased for a 5 year duration at a rate of $375/TB. No snapshots and backups provided.PIs can request snapshots and backups to a secondary cluster for an additional $375/TB (5 year duration).

HPC Research groups and Ceph

Research groups that purchase a minimum of 1TB Ceph project for a 5 year duration have the option of using their Ceph projects space as their home directory on Sol. PIs who opt for Ceph projects as home directory forgoing the 150GB home storage quota provided by the $50/user account fee will have their annual user fees waived for the life of the Ceph project. Such PIs can add users to their Sol allocation at no charge during the 5 year life of the Ceph projectEach Principle Investigator is provided with a 1TB ceph space for his/her research group. If additional space is requested, please include a justification in your compute allocation request or explicitly request a storage allocation. This storage exists as long as your allocation is active and will be deleted (no backups kept) a month after your allocation expires. If you purchase a Ceph project, then your storage will exist for  5 years irrespective of your compute allocation status.

Account Request: To obtain purchase a Ceph project, please contact Manager of Research Computing for more information.

...

Additional Storage: Additional storage is available by purchasing a Ceph project volume @ $375/TB for 5 years. To request additional home directory storage, please submit request to http://lts.lehigh.edu/help.
Definition of One Core-hour/Service Unit/SU: 1 hour of computing on 1 core. The base compute node on Sol with 20 cores will consume 20 SUs per hour of computing.

Anchor
teaching
teaching
Teaching uses of Research Computing Systems

Faculty members considering use of research computing facilities for teaching purposes should submit a request at https://lts.lehigh.edu/help at an Education Allocation at least eight weeks prior to the class start date with an anticipated enrollment count, a proposed syllabus, and details of their proposed use of HPC systems. 

These accounts are typically associated with a rostered course, and last for the duration of that course (up to one semester).  The course instructor can  Faculty should request these accounts for his/her students, and the department offering the course is responsible for paying the associated fees.  requests by TA's or department coordinators will be denied. Faculty are encouraged to submit a final report at the end of the semester to help us improve services provided for teaching.

A course allocation provides 1TB Ceph space and an allocation based on number of students in the course. The fee is broken down as follows

...

  • Provides 500 SU/student.
  • A course with 10 students will thus have a total allocation of 5000 SU that is shared among all students in the course.

...

Faculty can request storage only education allocations (please justify storage > 1TB).

Instructors requiring assistance with estimating total or per student SU requirements for the course should contact Research Computing Staff at least 4 weeks prior to beginning of the semester. 

Usage Policy: Student accounts cannot be shared and will be active until two weeks past the end of the semester. All compute intensive tasks must be submitted via the batch scheduler. We reserve the right to suspend accounts of users who run compute intensive jobs on the Sol login node. On request, LTS Research Computing staff will guest lecture on how to use the resource, write and submit job scripts and monitor jobs. Compute intensive task is defined as any operation on the HPC resource other than editing, copying, moving or deleting files, submitting and monitoring jobs and issuing simple commands such as ls, cp, mv, mkdir, rm, tail, tar, gzip/gunzip, more, cat and less. All student data not saved in the Ceph project space will be purged when accounts are deactivated.

Ceph Storage Resource

Faculty can request a Ceph project to provide storage resources for coursework independent of accounts on Research Computing clusters. The cost of a Ceph allocation is $200/TB/year and must be paid by the department offering the course. Any request for a Ceph volume for course work will need to be accompanied by

...

.