Welcome to the Network-Based Computing (NBC) Laboratory at the Computer Science and Engineering Department!
Many state-of-the-art and exciting research projects are being carried out by various members of the group, including:
- High Performance MPI on Infiniband Cluster
- Programming model support for Co-Processor (GPU & MIC)
- High Performance Computing with Virtualization
- Scalable Filesystems and I/O
- Performance Evaluation of Cluster Networking and I/O Technologies
- Networking, Storage and Middleware Support for Cloud Computing and DataCenters
- High Performance Runtime for PGAS Models (OpenSHMEM, UPC and CAF)
- BigData (Hadoop,Spark & Memcached)
The MVAPICH2 software, based on MPI 3.1 standard, delivers the best performance, scalability and fault tolerance for high-end computing systems and servers using InfiniBand, Omni-Path, Ethernet/iWARP, and RoCE networking technologies. This software is being used by more than 2,950 organizations in 86 countries worldwide to extract the potential of these emerging networking technologies for modern systems. As of Feb '19, more than 524,000 downloads have taken place from this project's site. This software is also being distributed by many vendors as part of their software distributions.
Multiple software libraries for Big Data processing and management, designed and developed by the group under High-Performance Big Data (HiBD) Project are available. These include: 1) RDMA-enabled Apache Hadoop Software library providing native RDMA (InfiniBand Verbs and RoCE) support for multiple components (HDFS, MapReduce and RPC) of Apache Hadoop; 2) RDMA-enabled Spark Software library providing native RDMA (InfiniBand Verbs and RoCE) support; 3) RDMA-Memcached Software library for providing native RDMA (InfiniBand Verbs and RoCE) support for Memcached used in Web 2.0 environment; and 4) OSU High-performance Big data Benchmarks (OHB). Sample performance numbers and download instructions for these packages are available from the above-mentioned website. These libraries are currently being used by more than 300 organizations worldwide in 30 countries. More than 29,150 downloads of this software have taken place from the project website alone.
The objectives of the research group are as follows:
- Proposing new designs for high performance network-based computing systems by taking advantages of modern networking technologies and computing systems
- Developing better middleware, API, and programming environments so that modern network-based computing applications can be developed and implemented in a scalable and high performance manner
- Performing the above research in an integrated manner (by taking systems, networking, and applications into account)
- Focusing on experimental computer science research
Multiple Positions available in the group:
1) Post-Doc/Research Scientist
2) MPI Software Engineer/Programmer
The projects in the Laboratory are funded by U.S. National Science Foundation, U.S. DOE Office of Science, U.S. Department of Defense, Ohio Board of Regents, Ohio Department of Development, Cisco Systems, Cray, Intel, Linux Networx, Mellanox, Microsoft, NVIDIA, Pattern Computer, QLogic, and Sun Microsystems; and equipment donations from Advanced Clustering, AMD, Appro, Chelsio, Dell, Fulcrum, Fujitsu, Intel, Mellanox, Microway, NetEffect, Pattern Computer, QLogic and Sun. Other technology partner includes: TotalView.
Upcoming Tutorials: InfiniBand, Omni-Path, and High-Speed Ethernet (HSE) at SC '18. Past tutorials: InfiniBand, Omni-Path, and High-Speed Ethernet (HSE) at ISC '18, HPC Meets Cloud at ICDCS '18, and MVAPICH2 and MPI-T at PEARC18.
Talk on TensorFlow Acceleration Presented at DataWorks Summit, Berlin, Germany.
Talk on Deep Learning over Big Data (DLoBD) presented at Spark + AI Summit
6th Annual MVAPICH User Group (MUG) Meeting took place on August 6-8, 2018 in Columbus, Ohio, USA. Click here for details.