AISys Lab. ECE, Seoul National University

Welcome to AISys!

Accelerated Intelligent Systems Lab (AISys) is affiliated with ECE, Seoul National University. We conduct research on system and architectural issues for accelerating various applications such as deep learning, compression algorithms and graph processing.

Hiring

AISys Lab is currently looking for talented students (graduate students, undergraduate interns). Please contact leejinho at snu dot ac dot kr if you are interested.

석박사 신입생 및 학부생 인턴을 상시 선발하고 있습니다. 관심있는 학생은 leejinho at snu dot ac dot kr 로 연락 바랍니다.

News

2026

Feb. 2026 Two papers have been accepted to DAC 2026. Warm congratulations to the authors!
  • TopGQ: Fast GNN Post-Training Quantization Leveraging Topology Information
  • SePArate: Segmenting Patterns from Defects in Wafer Manufacturing Using Weak Supervision
  • Feb. 2026 At HPCA 2026, our paper "LoCaLUT" was presented in the Best Paper Candidates session.
    Jan. 2026 Our paper GriNNder: Breaking the Memory Capacity Wall in Full-Graph GNN Training with Storage Offloading has been accepted to MLSys 2026. Congratulations to authors and see you at Bellevue!
    Jan. 2026 Kanghyun Choi, Hyeyoon Lee, Sunjong Park, and Dain Kwon won a gold prize, and Hongsun Jang, Jaeyong Song, Changmin Shin, and Si Ung Noh won a silver prize at the 32nd Samsung Humantech Paper Awards. Warm congratulations to the awardees!
    Jan. 2026 Our paper A Cost-Effective Near-Storage Processing Solution for Offline Inference of Long-Context LLMs has been accepted to ASPLOS 2026. Congratulations to authors and see you at Pittsburgh!

    2025

    Nov. 2025 Our paper Refined classification of YSOs and AGB stars by IR magnitudes, colors, and time-domain analysis with machine learning has been accepted to The Astrophysical Journal. Looks like our pursuit of system scalability reached stellar-system scale this time.
    Nov. 2025 Our paper FaScalSQL: A Fast and Scalable GPU-Accelerated SQL Query Engine for Out-of-Memory Tables has been accepted to ICDE 2026. Congratulations to authors! This is another work done in collaboration with HPCP Lab. (Prof. Youngsok Kim) at Yonsei Univ. and CISA Lab. (Prof. Joonsung Kim) at SKKU
    Nov. 2025 Our paper LoCaLUT: Harnessing Capacity-Computation Tradeoffs for LUT-Based Inference in DRAM-PIM has been accepted to HPCA 2026. Congratulations to authors! This work was done in collaboration with HPCP Lab. (Prof. Youngsok Kim) and Core Lab (Prof. Hanjun Kim) at Yonsei Univ.
    Sep. 2025 Our paper FALQON: Accelerating LoRA Fine-tuning with Low-Bit Floating-Point Arithmetic has been accepted to NeurIPS 2025. Congratulations to authors!
    Sep. 2025 Our paper FlexiWalker: Extensible GPU Framework for Efficient Dynamic Random Walks with Runtime Adaptation has been accepted to EuroSys. Congratulations!

    Research Topics

    We conduct research on system and architectural issues for accelerating various applications such as deep learning, compression algorithms and graph processing, especially on FPGAs and GPUs. Some of the on-going research topics are listed below. However, you’re free to bring your own exciting topic.

    AI Accelerators

    AI Accelerators

    With no doubt the most popular accelerator for AI nowadays is GPU. However the world is heading towards the next step: AI-specific accelerators. There is much room to improve in terms of accelerator designs. For example, optimizing dataflow, utilizing sparse network structure, or processing-in-memory techniques.

    Distributed Deep Learning

    Distributed Deep Learning

    To utilize multiple devices (i.e., GPUs) for high-speed DNN training, it’s common to employ distributed learning. There are still many ways to improve current distributed learning methods: Devising a new communication algorithm, smartly pipelining the jobs, or changing the ways that devices synchronize.

    Neural Network Acceleration and Compression

    Neural Network Acceleration and Compression

    Efficient deployment and training of deep neural networks require reducing both computation and memory overhead. Our research focuses on advanced quantization and compression techniques to accelerate both inference and training across a wide range of models, including large language models (LLMs), diffusion-based image generation, graph neural networks, and spiking neural networks. We aim to enable fast and resource-efficient learning and inference on various hardware platforms without significant loss of accuracy.

    System Optimization and Acceleration

    System Optimization and Acceleration

    We focus on optimizing and accelerating diverse programs and applications across emerging commercial hardwares including GPUs, CXL, and PIM (Processing-In-Memory). Our research spans multiple domains: communication libraries, graph processing algorithms, vector databases, and scientific applications. We optimize system software through a combination of hardware-aware acceleration and algorithmic optimization to achieve significant performance improvements for real-world workloads.