Research Overview

I study the problem of dataset distillation and its applications. The goal is to distill a large dataset into a small but information-rich synthetic dataset. We created the first large-scale benchmark(DC-BENCH) for the field and scaled up dataset distillation to ImageNet-1K.

Education

University of California, Los Angeles

Ph.D. in computer science

  • Graduate Student Researcher
  • Teaching Assistant

Carnegie Mellon University

M.S in Electrical & Computer Engineering

  • Selected Coursework - Machine Learning(10-701), Natual Language Processing(11-611), Advanced Cloud Computing(15-719), etc
  • GPA 3.98 / 4.0

Outreach

Internship I am looking for internship opportunities. My research focus is in dataset distillation and machine learning. I am open to all interesting research problems.

Selected Papers

Scaling Up Dataset Distillation to ImageNet-1K with Constant Memory

Justin Cui, Ruochen Wang, Si Si, Cho-Jui Hsieh
Paper Link

DC-BENCH Dataset Condensation Benchmark

Justin Cui, Ruochen Wang, Si Si, Cho-Jui Hsieh @ NeurIPS 2022
Paper Link

Community Services

Reviewer for NeurIPS 2022, ICML 2023