CCS Advanced Computing invites you to connect with the Advanced Computing community on Slack: http://umadvancedcomputing.slack.com
The Advanced Computing community Slack channels provide a place for user discussions, information sharing, and informal announcements about CCS resources and developments. All users with an @miami.edu or @umiami.edu email address can create a Slack account in UM Advanced Computing.
Files on /scratch are subject to purging after 21 days. Circumventing this purge is a violation of University of Miami Acceptable Use Policies and the account responsible for it will be suspended. Using ‘touch’ to thwart /scratch purge is not allowed.
CCS AC debuts the new ‘parallel’ Pegasus cluster job queue for parallel jobs requiring 16 cores from the same host, or MPI jobs requiring 16+ CPUs.
Click title for details.
The High-Performance Computing (HPC) core is focused on providing the latest in Supercomputing technology and tools to the University of Miami (UM) research community. While this core consists of traditional operations staff such as systems and network administrators, it also encompasses other areas of expertise including scientific programmers, parallel code profiling, and optimization. The HPC core is responsible for the operations of all infrastructure maintained at CCS.
The HPC core’s services include batch and interactive Compute, Visualization and Secure Data Processing clusters, Systems Administration and Consulting, Storage Implementation, Archive Storage, Systems Hosting and Maintenance. The HPC core has in-depth experience in parallelizing codes written in Fortran, C, Java, Perl, Python, and R. The core also has expertise in parallelizing code using both MPI and OpenMP.
Central to HPC is the Pegasus Supercomputer, a 350-node Lenovo cluster with over 300 applications and optimized libraries, including standard scientific libraries and numerous optimized libraries and algorithms tuned for the computing environment. The LSF (load sharing facility) resource manager, which supports over 1500 users and over 200,000 simultaneous job submissions, maximizes the efficiency of computational resources. By utilizing the full suite of LSF tools, we are able to provide for both batch and interactive jobs while retaining workload management features.