Chao Zhang

Table of Contents

Quick Links

James Edenfield Assistant Professor
School of Computational Science and Engineering
College of Computing
Georgia Institute of Technology

Office: CODA E1358B
Address: 756 W Peachtree St NW, Atlanta, GA 30308
Email: [email protected]

Research

My research lies in the areas of data science, machine learning, and AI. My goal is to make it easier to build domain-customized foundation models and AI agents for task-solving and decision-making. My technical efforts centers on addressing key challenges in data efficiency, computation efficiency, and model robustness. On the application front, I am deeply interested in harnessing foundation models to advance AI for science.

Currently, I am working on the following themes:

  1. Data-Centric LLM – Adapting Large Language Models for target domains by addressing data scarcity challenges through data-efficient methods such as learning from weak supervision and active learning.
  2. LLM Agents & Reasoning – Improving LLM reasoning and planning abilities by enabling them to learn and evolve through interaction with external environments for feedback. The goal is to better adapt LLMs and enhance their reasoning capabilities without expensive manual curation of fine-tuning data.
  3. AI Alignment – Ensuring responsible and reliable deployment of AI through critical techniques including uncertainty quantification, enhancing LLM factuality, and improving LLM alignment.
  4. AI for Science – Leveraging foundation models and AI agents to accelerate scientific discovery in diverse fields such as material science, biomedical and life sciences, and urban science.

Acknowledgment: My work has been generously supported by research funding/gift from NSF (IIS CAREER-2144338, IIS-2106961, IIS-2008334), ONR MURI , Kolon, HomeDepot, ADP, and Adobe. My work has also been recognized by an NSF CAREER Award, a Facebook Faculty Award, an Amazon AWS Machine Learning Research Award, a Google Faculty Research Award, a Kolon Faculty Fellowship, an ACM SIGKDD Dissertation Runner-up Award, and several paper awards from IMWUT (UbiComp), ECML/PKDD, and ML4H.

I. Data-Centric LLM

We aim to adapt LLMs to various domains and complex tasks. A key bottleneck when adapting LLMs is data scarcity – the lack of high-quality, representative data for the target domain. We are tackling this bottleneck by developing methods for data-efficient LLM adaptation, including:

II. LLM Agents

We investigate improving LLM reasoning and planning abilities by enabling them to evolve through interaction with external environments for feedback. The goal is to better adapt LLMs and enhance their reasoning capabilities without expensive manual curation of fine-tuning data.

III. AI Alignment

We aim to develop AI systems that are not only capable but also trustworthy for deployment in various domains. We study the following topics in the space of trustworthy AI: uncertainty quantification, LLM factuality, and LLM alignment.

IV. AI for Science

We aim to leverage AI and foundation models for advancing scientific discovery. We develop domain-specific foundation models and LLM agents for different scientific domains. On the application side, we collaborate with domain-experts to advance scientific discovery in material design, biomedical and life science, and urban science:

Awards

  • 2024 GaTech CoC Outstanding Junior Faculty Award
  • 2022 NSF Career Award
  • 2022 ML4H Outstanding Paper Award
  • 2021 Facebook Faculty Research Award
  • 2021 Kolon Faculty Fellowship
  • 2020 Amazon AWS Machine Learning Research Award
  • 2020 Google Faculty Research Award
  • 2019 ACM SIGKDD Dissertation Award Runner-up
  • 2018 ACM IMWUT Distinguished Paper Award
  • 2015 ECML/PKDD Best Student Paper Runner-up Award
  • 2013 Chiang Chen Overseas Graduate Fellowship

Publications

(* denotes equal contribution)

2025

2024

2023

2022

2021

2020

2019

2018

Earlier

Teaching

Students

Prospective students: I am always looking for strong and motivated students to join our group. If you are interested in working with me, you can either email me or fill out this form.

Current:

  • Rui Feng: Ph.D. Student in CS
  • Yuchen Zhuang: Ph.D. Student in ML
  • Yinghao Li: Ph.D. Student in ML
  • Rongzhi Zhang: Ph.D. Student in ML
  • Haotian Sun: Ph.D. Student in ML (co-advised with Bo Dai)
  • Kuan Wang: Ph.D. Student in CSE
  • Haorui Wang: Ph.D. Student in CSE
  • Agam A. Shah: Ph.D. Student in ML (co-advised with Sudheer Chava)
  • Rushi Qiang: Ph.D. Student in CSE (co-advised with Bo Dai)
  • Changhao Li: Ph.D. Student in CSE (co-advised with Bo Dai)

Alumni:

  • Yue Yu: Ph.D., 2024 (–> Research Scientist @ Meta GenAI Team)
  • Lingkai Kong: Ph.D., 2024 (–> Postdoc Fellow @ Harvard)
  • Yanbo Xu: Ph.D., 2023 (–> Research Scientist @ Microsoft Research)
  • Binghong Chen: Ph.D., 2023 (–> Quant @ Citadel Capital, co-advised with Prof. Le Song)
  • Pranav Shetty: Ph.D., 2023 (–> Research Scienctist @ JP Morgan Chase, JP Morgan AI Ph.D. Fellowship, co-advised with Prof. Rampi Ramprasad)
  • Yi Rong: Visiting Ph.D. Student
  • Vidit Jain: M.S. Student in CS
  • Mukund Rungta: M.S. Student in CS
  • Junyang Zhang: M.S. Student in CS
  • Piyush Patil: M.S. Student in CS
  • Mengyang Liu: M.S. Student in CSE
  • Isaac Rehg: M.S. in CS
  • Wendi Ren: M.S. in CSE
  • Ruijia Wang: M.S. in CSE
  • Jacob Wessell: M.S. in CS
  • Wenhao Mu: M.S. in CS
  • Shangqing Xu: M.S. in CS