Chao Zhang

Table of Contents

Quick Links

Assistant Professor
School of Computational Science and Engineering
College of Computing
Georgia Institute of Technology

Office: CODA E1358B
Address: 756 W Peachtree St NW, Atlanta, GA 30308
Email: chaozhang@gatech.edu

Research

I develop data science and machine learning models to assist with task solving and decision-making. My research focuses on rapidly developing AI systems that are more data-efficient, trustworthy, and customized for specific domains.

Currently, I am working on the following themes:

  1. Data-Efficient AI – Adapting Large Language Models for target domains by addressing data scarcity challenges through data-efficient methods such as feedback-based adaptation, weak supervision, and active learning.
  2. Trustworthy AI – Making AI models more reliable and aligned with human values through uncertainty quantification, robust fine-tuning, scalable alignment, and decision-focused learning.
  3. AI for Science – Leveraging data science and AI to accelerate scientific discovery in fields like material science, biomedical and life sciences, and urban science through close collaboration with domain experts.

Acknowledgment: My work has been generously supported by research funding/gift from NSF (IIS CAREER-2144338, IIS-2106961, IIS-2008334), ONR MURI , Kolon, HomeDepot, ADP, and Adobe. My work has also been recognized by an NSF CAREER Award, a Facebook Faculty Award, an Amazon AWS Machine Learning Research Award, a Google Faculty Research Award, a Kolon Faculty Fellowship, an ACM SIGKDD Dissertation Runner-up Award, and several paper awards from IMWUT (UbiComp), ECML/PKDD, and ML4H.

I. Data-Efficient AI

We aim to adapt LLMs to various domains and complex tasks. A key bottleneck when adapting LLMs is data scarcity – the lack of high-quality, representative data for the target domain. We are tackling this bottleneck by developing methods for data-efficient LLM adaptation, including:

II. Trustworthy AI

We aim to develop AI systems that are not only capable but also trustworthy for deployment in various domains. We focus on several aspects of trustworthy AI: uncertainty quantification for deep learning, robust LLM fine-tuning and alignment, and decision-focused learning.

III. AI for Science

We aim to leverage data science and AI for scientific applications ranging from macro-level systems like transportation and epidemiology to micro-level systems such as molecules. We have been developing data-driven tools to advance scientific discovery in material design, biomedical and life science, and urban science:

Awards

  • 2024 GaTech CoC Outstanding Junior Faculty Award
  • 2022 NSF Career Award
  • 2022 ML4H Outstanding Paper Award
  • 2021 Facebook Faculty Research Award
  • 2021 Kolon Faculty Fellowship
  • 2020 Amazon AWS Machine Learning Research Award
  • 2020 Google Faculty Research Award
  • 2019 ACM SIGKDD Dissertation Award Runner-up
  • 2018 ACM IMWUT Distinguished Paper Award
  • 2015 ECML/PKDD Best Student Paper Runner-up Award
  • 2013 Chiang Chen Overseas Graduate Fellowship

Publications

(* denotes equal contribution)

2025

  • Hephaestus: Improving Fundamental Agent Capabilities of Large Language Models through Continual Pre-Training
    Yuchen Zhuang, Jingfeng Yang, Haoming Jiang, Xin Liu, Kewei Cheng, Sanket Lokegaonkar, Yifan Gao, Qing Ping, Tianyi Liu, Binxuan Huang, Zheng Li, Zhengyang Wang, Pei Chen, Ruijie Wang, Rongzhi Zhang, Nasser Zalmout, Priyanka Nigam, Bing Yin, Chao Zhang
    Annual Conference of the North American Chapter of the Association for Computational Linguistics (NAACL), 2025

2024

2023

2022

2021

2020

2019

2018

Earlier

Teaching

Students

Prospective students: I am always looking for strong and motivated students to join our group. If you are interested in working with me, you can either email me or fill out this form.

Current:

  • Yue Yu: Ph.D. Student in CSE
  • Rui Feng: Ph.D. Student in CS
  • Yuchen Zhuang: Ph.D. Student in ML
  • Yinghao Li: Ph.D. Student in ML
  • Rongzhi Zhang: Ph.D. Student in ML
  • Haotian Sun: Ph.D. Student in ML (co-advised with Bo Dai)
  • Kuan Wang: Ph.D. Student in CSE
  • Haorui Wang: Ph.D. Student in CSE
  • Agam A. Shah: Ph.D. Student in ML (co-advised with Sudheer Chava)
  • Rushi Qiang: Ph.D. Student in CSE (co-advised with Bo Dai)
  • Changhao Li: Ph.D. Student in CSE (co-advised with Bo Dai)
  • Jacob Wessell: M.S. student in CS
  • Wenhao Mu: M.S. student in CS
  • Shangqing Xu: M.S. student in CS

Alumni:

  • Lingkai Kong: Ph.D., 2024 (–> Postdoc Fellow @ Harvard)
  • Yanbo Xu: Ph.D., 2023 (–> Research Scienctist @ Microsoft Research)
  • Binghong Chen: Ph.D., 2023 (–> Quant @ Citadel Capital, co-advised with Prof. Le Song)
  • Pranav Shetty: Ph.D., 2023 (–> Research Scienctist @ JP Morgan Chase, JP Morgan AI Ph.D. Fellowship, co-advised with Prof. Rampi Ramprasad)
  • Vidit Jain: M.S. Student in CS
  • Mukund Rungta: M.S. Student in CS
  • Junyang Zhang: M.S. Student in CS
  • Piyush Patil: M.S. Student in CS
  • Mengyang Liu: M.S. Student in CSE
  • Isaac Rehg: M.S. in CS
  • Wendi Ren: M.S. in CSE
  • Ruijia Wang: M.S. in CSE
  • Yi Rong: Visiting Ph.D. Student