Skip to Main Content

High Performance Computing @ UMBC

The following high performance computers have been operated by the UMBC High Performance Computing Facility:

  • maya.rs.umbc.edu is the current cluster of the UMBC High Performance Computing Facility (HPCF). It is a heterogeneous cluster composed of equipment purchased in 2009 through 2013 and released to the user community in Spring 2014. It contains several groups of nodes with an InfiniBand interconnect:
    • 34 Dell PowerEdge R620 CPU-only compute nodes. Each node has two Intel Ivy Bridge 2650v2 (2.6 GHz, 20 MB cache), processors with eight cores apiece, for a total of 16 cores per node.
    • 19 Dell PowerEdge R720 CPU/GPU nodes, each with the above CPUs plus two (2) NVIDIA Tesla K20 GPUs.
    • 19 Dell PowerEdge R720 CPU/Phi nodes, each with the above CPUs plus two (2) Phi 5110P coprocessors.
    • 168 IBM iDataPlex CPU nodes featuring two (2) quad core Intel Nehalem X5560 processors (2.8 GHz, 8 MB cache) and 24 GB of memory.
    • 86 IBM iDataPlex CPU-only nodes featuring two (2) quad core Intel Nehalem X5550 processors (2.66 GHz, 8 MB cache) and 24 GB of memory.

    For a detailed description and more information see resources for maya users. The purchase of maya equipment was supported by faculty contributions, by grants from the National Science Foundation, and by the UMBC administration. Researchers can purchase nodes for long-term priority access. If you are interested in joining the effort, see the contact page.

  • tara.rs.umbc.edu is the current cluster in the UMBC High Performance Computing Facility (HPCF). Installed in Fall 2009, tara is an 86-node distributed-memory cluster with two quad-core Intel Nehalem processors and 24 GB per node, an InfiniBand interconnect, and 160 TB central storage. For a detailed description and more information see 2013 resources for tara users or 2010 resources for tara users. The purchase of tara was supported by faculty contributions, by grants from the National Science Foundation, and by the UMBC administration.
  • hpc.rs.umbc.edu was a distributed-memory cluster with 33 compute nodes plus 1 development and 1 combined user/management node, each equipped with two dual-core AMD Opteron processors and at least 13 GB of memory, connected by an InfiniBand network and with an Infiniband-accessible 14 TB parallel file system. The purchase in 2008 pooled funds from several researchers with seed funding from UMBC. This machine was replaced by tara in 2009. For the historical record of hpc and the projects that used it, see thearchived hpc webpage.
  • kali.math.umbc.edu was a distributed-memory cluster with 32 compute nodes including a storage node connecting the 0.5 TB central storage, each with two Intel Xeon 2.0 GHz processors and 1 GB of memory, connected by a Myrinet interconnect, plus 1 combined user/management node. This machine was purchased in 2003 by researchers from the Department of Mathematics and Statistics using a grant from the National Science Foundation. It became part of HPCF in 2008 and was shut down in early 2009 after the cluster hpc had become operational. For the historical record of kali and the projects that used it, see the kali webpage.

Other high performance computers available on the UMBC campus are:

  • bluegrit.cs.umbc.edu is a distributed-memory cluster with 47 blades using IBM PowerPC chips (33 blades with 0.5 GB memory and 14 dual-processor blades with 2 GB memory) and 12 blades with two IBM Cell Processors and 0.5 GB memory. See the bluegrit webpage for more information. This machine is part of the Multicore Computational Center (MC2) at UMBC.