GAUDI Lab General-Purpose Architectures with Unleashed Design Innovations

Research

Research Topics

Our research topics include, but are not limited to, the following areas:

feature image

Computer System Architectures and Resource Orchestration

Modern computer systems have become larger and more complex, underscoring the need for designing highly efficient system architectures and effectively orchestrating shared resources, such as memory and Last-Level Cache (LLC). We are investigating cutting-edge systems and exploring diverse solutions to improve system-wide processing efficiency.

feature image

CPU Core Microarchitectures and Multi-core Architectures

The Central Processing Unit (CPU) plays a vital role in a system by running applications and system software, managing system resources, interacting with I/O devices, etc. We are exploring advanced microarchitectural techniques, such as dynamic scheduling and speculation, as well as innovative multi-core architectures like heterogeneous multi-core and simultaneous multi-threading, to enhance the CPU’s performance and energy efficiency.

feature image

GPGPU and Accelerator Architectures

Emerging applications with diverse computation patterns have driven the development of specialized hardware designed to deliver optimal performance per watt for these tasks. We are exploring energy-efficient architectures for General-Purpose Graphics Processing Units (GPGPUs) and Neural Processing Units (NPUs). In addition, we are investigating various use cases for on-chip accelerators, which can offload specific tasks from the CPU to specialized hardware units to alleviate datacenter/system taxes.

feature image

Near-Data Processing

Data-centric applications encounter significant bottlenecks in transferring large volumes of data between processors and memory, storage, or networks. To address these challenges, we are investigating efficient near-data processing techniques at various system layers. This includes Processing-In-Memory (PIM), integrating processing capabilities directly into memory; In-Storage Processing (ISP), enabling data processing within storage devices; and In-Network Computing (INC), performing computations within the network infrastructure.

feature image

Emerging Interconnect Technologies

Compute eXpress Link (CXL) is an open standard for high-speed, efficient interconnects between CPUs and other devices such as memory, accelerators (e.g., GPUs, FPGAs), and SmartNICs (Network Interface Cards). It is designed to enhance performance and resource sharing in data centers, particularly in large-scale computing. We are exploring various use cases for CXL technology, including its potential to enable more flexible memory expansion, improve accelerator integration, and optimize data flow between heterogeneous computing components in next-generation data center architectures.

Collaboration

**Future Architecture and System Technology for Scalable Computing (FAST)** <br> University of Illinois Urbana-Champaign (UIUC), IL, United States
Future Architecture and System Technology for Scalable Computing (FAST)
University of Illinois Urbana-Champaign (UIUC), IL, United States
**Embedded Systems and Computer Architecture Lab (eSCaL)** <br> Yonsei University, Seoul, Republic of Korea
Embedded Systems and Computer Architecture Lab (eSCaL)
Yonsei University, Seoul, Republic of Korea