Home » DiCoS Core » Computational Resources

Computational Resources

UD’s IT Research Computing team maintains several community clusters, co-owned by faculty members or groups. DSI owns a number of nodes on the most recently deployed “Caviness” cluster and has access to the new DARWIN cluster. UD Faculty members can request access to the DSI owned nodes for their research groups through DSI (see below).

DSI Sponsored Compute Resources

DARWIN Compute and Storage System

The DARWIN system is a new system currently available to early access users with startup allocation requests starting on February 16, 2021. It is designed to support research with diverse technology requirements often found in interdisciplinary sciences such as high I/O and low latency, scalability, GPU enabled execution, visualization acceleration, high memory workloads, and large local scratch space.

For more information about what kinds of allocations can be requested and how to apply please see the Allocation Information/Guidelines. Startup allocations request are currently being collected via this form (requires udel.edu login). All users agree to abide by the Acceptable Usage Agreement (internal users) and XSEDE Usage Policy (external users). For questions/problems with allocation for DSI sponsored compute resources, send an email to dsiallocations@udel.edu.

Caviness Community Cluster

The Caviness cluster, UD’s third Community High-Performance Computing (HPC) Cluster, is a distributed-memory Linux cluster. The first-generation of Caviness consists of containing 194 compute nodes aggregating to 7,336 Intel CPU cores; 110,080 CUDA cores; 4,800 Turing tensor cores; 46TiB RAM; 450TiB high-speed lustre scratch; and 225TiB NFS workgroup long-term storage. As a Caviness stakeholder, DSI owns two nodes with the following specification: 2 * 20C Intel Zeon Gold 6230; 768GiB RAM; 750GiB local scratch; nVidia T4 GPU; OmniPath high-speed networking.

To request an allocation from Caveness please use the Allocation Request Application. All users agree to abide by the Acceptable Usage Agreement. For questions/problems with allocation for DSI sponsored compute resources, send an email to dsiallocations@udel.edu.

Other UD Compute Resources

BioMix High-Performance Computing (HPC) Cluster

BioMix is a special purpose HPC system geared toward bioinformatics applications. It has >600 CPU cores with a combined 3.5TB of RAM and 372TB of disk space to support bioinformatics analysis. Included are 425 CPU cores which are made freely available to researchers as part of our Linux-based BioMix cluster, where all nodes are connected via 1 or 10 gigabit Ethernet and have access to a 20TB shared RAID array. A combination of system types allows the choice of systems best suited for the particular needs of a given analysis. These include numerous nodes configured for memory intensive computing applications with 128 – 512GB of RAM per machine.

Farber High-Performance Computing (HPC) Cluster

Farber is a predecessor HPC system to Caviness. It’s a distributed-memory, Linux cluster consisting of 100 compute nodes, which total 2000 “Ivy Bridge” cores and 6.4 TB RAM. Storage options include a 288TB Lustre filesystem for on-line processing, and a 72TB NFS filesystem for workgroup and long-term storage. The cluster is located in the University’s core data center; it is supported by the central IT department, with redundant power, cooling, and network connectivity. It has a 56Gbps (FDR) InfiniBand network for MPI and Luster fast disk access, a 1Gbps TCP/IP network for scheduling and NFS, and 2 10Gbps links to the Core UD network.

National/External Resources

The National Science Foundation (NSF)’s XSEDE project (xsede.org) facilitates access to NSF supported computational resources. Access is free to projects of all funding sources (with some preference given to NSF projects). A startup location can be requested through a short proposal at any time. Larger allocations require a proposal to the XSEDE Resource Allocation Committee (XRAC), with quarter-annual submission deadlines – see portal.xsede.org/allocations. Proposals can also include requests for time of computational experts.

The Department of Energy (DOE)’s INCITE program makes available high-performance computing resources to researchers conducting projects in DOE’s mission, which is fairly broad. Proposals are accepted annually with a deadline typically in June – see www.doeleadershipcomputing.org

Other agencies may also offer access to their computing and data resources for projects in their mission. Examples:

Cloud Computing resources

Under development: DSI is considering obtaining an allocation of cloud computing resources for use by UD faculty members for DSI-related projects. We’d like to hear from UD members who have experience obtaining and using such resources.  Please contact us at dsi-core@udel.edu.

Our Mission

The Institute aims to accelerate research in data science, serving as a nucleating effort to catalyze interdisciplinary research collaborations across fields impacting our society.