Profile picture

Yue Zhao

Assistant Professor of Computer Science

Education

  • 2023, Doctoral Degree, Information Systems, Carnegie-Mellon University
  • 2016, Master's Degree, Computer Science, University of Toronto
  • 2015, Bachelor's Degree, Computer Engineering, University of Cincinnati

Biography

Dr. Yue Zhao is an Assistant Professor of Computer Science at the University of Southern California, where he leads the FORTIS Lab. His research focuses on AI auditing — building methods, benchmarks, and open-source tools that make AI systems inspectable, safe, and accountable. Dr. Zhao has authored over 70 peer-reviewed papers in top-tier venues and is internationally recognized for his open-source contributions, including PyOD, PyGOD, agent-audit, Aegis, and TrustLLM, which collectively exceed 35 million downloads and 28,000 GitHub stars. His tools are widely used across academia, industry, and government, including by NASA, Tesla, Morgan Stanley, and the U.S. Senate. He has received numerous honors, including the NVIDIA Academic Grant Program Award, the Capital One Research Award, multiple Amazon Research Awards, AAAI New Faculty Highlights, Google Cloud Research Innovators, the Norton Labs Fellowship, the Meta AI4AI Research Award, the Carnegie Mellon University Presidential Fellowship, the 2025 SIGSPATIAL Best Short Paper Award, and the Second Prize CCC Award at the IEEE ICDM 2025 BlueSky Track. He serves as an Associate or Action Editor for ACM Transactions on AI for Science, IEEE Transactions on Neural Networks and Learning Systems, and the Journal of Data-Centric Machine Learning Research, and as an Area Chair for major conferences including ICLR, ICML, and NeurIPS.

Research Summary

My research focuses on AI auditing, safety, and security for agent systems, with related work on foundation models and their deployment in real-world environments. I study how modern AI systems fail, how their behavior and risks can be evaluated, monitored, and controlled at the system level, and how they can be deployed responsibly in domains where failures carry significant consequences. This agenda is organized around three closely connected directions:

1. AI Auditing & Assurance
I develop methods, benchmarks, and open-source systems for auditing foundation models and agent systems. This work includes trustworthiness evaluation, security analysis of AI pipelines, ecosystem-scale risk scanning, and continuous monitoring that provides evidence for assurance in deployment. Representative systems include agent-audit and TrustLLM.

Keywords: AI Auditing, AI Assurance, Trustworthy AI, Foundation Models, Agent Systems, TrustLLM, agent-audit, Monitoring, Evaluation Frameworks, Risk Analysis

2. AI Safety & Security
I study failure modes, attack surfaces, and runtime control layers in large language models and agent systems, and design methods to detect and mitigate unsafe or compromised behavior. Representative topics include hallucinations, jailbreaks, prompt attacks, privacy leakage, model extraction, failures in multi-agent interactions, and runtime guardrails such as policy enforcement and approval workflows for tool-using agents. Representative systems include Aegis. This direction is also informed by my earlier work on anomaly detection and out-of-distribution detection.

Keywords: LLM Safety, Agent Safety, AI Security, AI Agent Security, Runtime Guardrails, Policy Enforcement, Hallucination Mitigation, Jailbreak Detection, Prompt Attacks, Privacy Leakage, Model Extraction, Robustness, OOD Detection, Anomaly Detection

3. AI for Science & Society
I apply reliable and auditable AI systems to domains where correctness, safety, and accountability matter, including climate and weather forecasting, healthcare and biomedicine, and computational social systems. These applications also serve as demanding testbeds for auditing, assurance, and safety methods.

Keywords: AI for Science, Climate AI, Weather Forecasting, Healthcare AI, Biomedicine, Computational Social Systems, Decision Modeling, High-Stakes AI

Awards

  • 2026 Amazon Amazon Research Awards
  • 2026 Nvidia NVIDIA Academic Grant Program
  • 2025 ACM SIGSPATIAL Best Short Paper
  • 2025 IEEE ICDM BlueSky Track Second Prize CCC Award
  • 2024 Amazon Amazon Research Awards
  • 2024 Google Google Cloud Research Innovators
  • 2024 Capital One Research Awards
  • 2024 Association for the Advancement of Artificial Intelligence AAAI New Faculty Highlights
Appointments
  • Thomas Lord Department of Computer Science
Office
  • Yue Zhao has not listed an office location.
Contact Information
  • yue.z@usc.edu
Links
Social Media