About PI

Professor Zhangyang “Atlas” Wang [Google Scholar] is a tenured Associate Professor and holds the Temple Foundation Endowed Faculty Fellowship #7, in the Chandra Family Department of Electrical and Computer Engineering at The University of Texas at Austin. He is also a faculty member of UT Computer Science (GSC) [CSRankings], and the Oden Institute CSEM program. Since May 2024, Dr. Wang has been on leave from UT Austin to serve as the full-time Research Director for XTX Markets, heading their new AI Lab in New York City.

Previously, he was the Jack Kilby/Texas Instruments Endowed Assistant Professor in the same department from 2020 to 2023; and an Assistant Professor of Computer Science and Engineering at Texas A&M University from 2017 to 2020. Alongside his academic career, he has also explored a few exciting opportunities in the industry. He was a visiting scholar at Amazon Search from 2021 to 2022; and later became the part-time Director of AI Research & Technology for Picsart from 2022 to 2024, leading the development of cutting-edge GenAI algorithms for visual creation and editing. He earned his Ph.D. in Electrical and Computer Engineering from UIUC in 2016, under the guidance of Professor Thomas S. Huang, and his B.E. in EEIS from USTC in 2012.

Prof. Wang has broad research interests spanning from the theory to the application aspects of machine learning (ML). At present, his core research mission is to leverage, understand and expand the role of low dimensionality in ML and optimization, whose impacts span over many important topics such as the efficiency and trust issues in large language models (LLMs) as well as generative vision. His research is gratefully supported by NSF, DARPA, ARL, ARO, IARPA, DOE, as well as dozens of industry and university grants. Prof. Wang co-founded the new Conference on Parsimony and Learning (CPAL) and served as its inaugural Program Chair. He is an elected technical committee member of IEEE MLSP and IEEE CI; and regularly serves as (senior) area chairs, invited speakers, tutorial/workshop organizers, various panelist positions and reviewers. He is an ACM Distinguished Speaker and an IEEE senior member.

Prof. Wang has received many research awards, including an NSF CAREER Award, an ARO Young Investigator Award, an IEEE AI's 10 To Watch Award, an AI 100 Top Thought Leader Award, an INNS Aharon Katzir Young Investigator Award, a Google Research Scholar award, an IBM Faculty Research Award, a J. P. Morgan Faculty Research Award, an Amazon Research Award, an Adobe Data Science Research Award, a Meta Reality Labs Research Award, and two Google TensorFlow Model Garden Awards. His team has won the Best Paper Award at the inaugural Learning on Graphs (LoG) Conference 2022, the Best Paper Finalist Award at the International Conference on Very Large Databases (VLDB) 2024, and five research competition prizes at CVPR/ICCV/ECCV since 2018. He feels most proud of being surrounded by some of the world's most brilliant students: his Ph.D. students include winners of seven prestigious fellowships (NSF GRFP, IBM, Apple, Adobe, Amazon, Qualcomm, and Snap), among many other honors.

About Our Research

At the VITA group, we pursue cutting-edge research spanning the theoretical foundations to practical applications of machine learning (ML). Our group's research continues to evolve, embracing new challenges at the forefront of AI and ML. We collaborate closely with industry partners and other academic institutions to ensure our work has real-world impact and addresses pressing technological needs.

Our current work is organized around three key themes, throughout which we maintain a commitment to developing ML algorithms that are efficient, scalable, and robust. We also explore the broader implications of our work, including applications in robotics, healthcare, and AI for social good.

Theme 1: Large Language Models (LLMs) - Fundamental Optimization, Mechanistic Understanding, & System Co-design

We focus on advancing the efficiency, scalability and trust of LLMs through innovative approaches to training and inference. Our research explores memory-efficient LLM training techniques (GaLoRe & LiGO), efficient generative inference methods (H2O & Flextron), understanding pre-trained model weights (essential sparsity & lottery ticket) or training artifacts (oversmoothening & LLM-PBE): many accompanied with system or hardware co-design.

Selected Notable Works:
  • J. Zhao, Z. Zhang*, B. Chen, Z. Wang, A. Anandkumar, and Y. Tian, "GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection”, International Conference on Machine Learning (ICML), 2024. (Oral) [Paper] [Code] (Integrated into HuggingFace, PyTorch, LLaMA-Factory, FedML, Axolotl, etc.)
  • R. Cai*, S. Muralidharan, G. Heinrich, H. Yin, Z. Wang, J. Kautz, and P. Molchanov, “Flextron: Many-in-One Flexible Large Language Model”, International Conference on Machine Learning (ICML), 2024. (Oral) [Paper] [Code]
  • Q. Li, J. Hong*, C. Xie, J. Tan, R. Xin, J. Hou, X. Yin, Z. Wang, D. Hendrycks, Z. Wang, B. Li, B. He, and D. Song, “LLM-PBE: Assessing Data Privacy in Large Language Models”, International Conference on Very Large Data Bases (VLDB), 2024. (Best Paper Finalist) [Paper] [Code]
  • Z. Zhang*, Y. Sheng, T. Zhou, T. Chen*, L. Zheng, R. Cai*, Z. Song, Y. Tian, C. Ré, C. Barrett, Z. Wang, and B. Chen, "H2O: Heavy-Hitter Oracle for Efficient Generative Inference of Large Language Models”, Advances in Neural Information Processing Systems (NeurIPS), 2023. [Paper] [Code] (Integrated into DeepSpeed by Microsoft, and Llama-Recipes by Meta)
  • A. Jaiswal*, S. Liu*, T. Chen*, Z Wang, "The Emergence of Essential Sparsity in Large Pre-trained Models: The Weights that Matter”, Advances in Neural Information Processing Systems (NeurIPS), 2023. [Paper] [Code]
  • P. Wang*, R. Panda, L. Hennigen, P. Greengard, L. Karlinsky, R. Feris, D. Cox, Z. Wang, and Y. Kim, "Learning to Grow Pretrained Models for Efficient Transformer Training”, International Conference on Learning Representations (ICLR), 2023. (Spotlight) [Paper] [Code] (Implemented in IBM’s AI production system)
  • P. Wang*, W. Zheng*, T. Chen*, and Z. Wang, “Anti-Oversmoothing in Deep Vision Transformers via the Fourier Domain Analysis: From Theory to Practice”, International Conference on Learning Representations (ICLR), 2022. [Paper] [Code]
  • T. Chen*, J. Frankle, S. Chang, S. Liu, Y. Zhang, Z. Wang, and M. Carbin, “The Lottery Ticket Hypothesis for Pre-trained BERT Networks”, Advances in Neural Information Processing Systems (NeurIPS), 2020. [Paper] [Code]

  • Theme 2: Optimization in Modern ML - Learning to Optimize, Black-box optimization, Neurosymbolic learning

    Our research in this theme focuses on developing novel optimization techniques for modern machine learning challenges. We have spearheaded the Learning to Optimize (L2O) framework (LISTA-CPSS, ALISTA, & HyperLISTA) and benchmark (L2O Primer), and recently explore the new frontiers in black-box LLM optimization (DP-OPT) and neurosymbolic AI (formal fine-tuning, symbolic L2O, & symbolic visual RL).

    Selected Notable Works:
  • W. Zheng*, S. Sharan*, Z. Fan*, K. Wang*, Y. Xi*, and Z. Wang, “Symbolic Visual Reinforcement Learning: A Scalable Framework with Object-Level Abstraction and Differentiable Expression Search”, IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2024. [Paper] [Code]
  • Y. Yang, N. Bhatt*, T. Ingebrand, W. Ward, S. Carr, Z. Wang, and U. Topcu, "Fine-Tuning Language Models Using Formal Methods Feedback”, Conference on Machine Learning and Systems (MLSys), 2024. [Paper] [Code]
  • J. Hong*, J. Wang, C. Zhang, Z. LI*, B. Li, and Z. Wang, "DP-OPT: Make Large Language Model Your Differentially-Private Prompt Engineer”, International Conference on Learning Representations (ICLR), 2024. (Spotlight) [Paper] [Code]
  • (α-β) T. Chen*, X. Chen*, W. Chen*, H. Heaton, J. Liu, Z. Wang, and W. Yin, “Learning to Optimize: A Primer and A Benchmark”, Journal of Machine Learning Research (JMLR), 2022. [Paper] [Code]
  • W. Zheng*, T. Chen*, T. Hu*, and Z. Wang, “Symbolic Learning to Optimize: Towards Interpretability and Scalability”, International Conference on Learning Representations (ICLR), 2022. [Paper] [Code]
  • X. Chen*, J. Liu, Z. Wang, W. Yin, “Hyperparameter Tuning is All You Need for LISTA”, Advances in Neural Information Processing Systems (NeurIPS), 2021. [Paper] [Code]
  • J. Liu, X. Chen*, Z. Wang, and W. Yin, “ALISTA: Analytic Weights Are As Good As Learned Weights in LISTA”, International Conference on Learning Representations (ICLR), 2019. [Paper] [Code]
  • X. Chen*, J. Liu, Z. Wang, and W. Yin, “Theoretical Linear Convergence of Unfolded ISTA and Its Practical Weights and Thresholds”, Advances in Neural Information Processing Systems (NeurIPS), 2018. (Spotlight) [Paper] [Code]

  • Theme 3: Generative Vision - 3D/4D/Video Synthesis, and Related Applications

    Our group's earlier (pre-2021) work includes several influential algorithms for GAN-based image enhancement and editing “in the wild”. More recently (post-2021), we push the boundaries of generative AI for visual tasks, with a focus on 3D/4D reconstruction (LSM, InstantSplat, LightGaussian, & NeuralLift-360), novel view synthesis (GNT & SinNeRF), and video generation (StreamingT2V & Text2Video-Zero).

    Selected Notable Works:
  • Z. Fan*, J. Zhang, W. Cong*, P. Wang*, R. Li, K. Wen, S. Zhou, A Kadambi, Z. Wang, D. Xu, B. Ivanovic, M. Pavone, and Y. Wang, “Large Spatial Model: End-to-end Unposed Images to Semantic 3D”, Advances in Neural Information Processing Systems (NeurIPS), 2024. [Paper] [Code]
  • Z. Fan*, K. Wang*, K. Wen, Z. Zhu*, D. Xu*, and Z. Wang, "LightGaussian: Unbounded 3D Gaussian Compression with 15x Reduction and 200+ FPS”, Advances in Neural Information Processing Systems (NeurIPS), 2024. (Spotlight) [Paper] [Code]
  • Z. Fan*, W. Cong*, K. Wen, K. Wang*, J. Zhang, X. Ding, D. Xu, B. Ivanovic, M. Pavone, G. Pavlakos, Z. Wang, and Y. Wang, "InstantSplat: Unbounded Sparse-view Pose-free Gaussian Splatting in 40 Seconds”, arXiv preprint arXiv:2403.20309, 2024. [Paper] [Code]
  • R. Henschel, L. Khachatryan, D. Hayrapetyan, H. Poghosyan, Vahram Tadevosyan, V. Tadevosyan, Z. Wang, S. Navasardyan, and H. Shi, "StreamingT2V: Consistent, Dynamic, and Extendable Long Video Generation from Text”, arXiv preprint arXiv:2403.14773, 2024. [Paper] [Code]
  • L. Khachatryan, A. Movsisyan, V. Tadevosyan, R. Henschel, Z. Wang, S. Navasardyan, and H. Shi, "Text2Video-Zero: Text-to-Image Diffusion Models are Zero-Shot Video Generators”, IEEE International Conference on Computer Vision (ICCV), 2023. (Oral) [Paper] [Code] (Commercialized as Picsart AI GIF generator)
  • D. Xu*, Y. Jiang*, P. Wang*, Z. Fan*, Y. Wang*, and Z. Wang, "NeuralLift-360: Lifting An In-the-wild 2D Photo to A 3D Object with 360◦ Views”, IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2023. (Highlight) [Paper] [Code]
  • M. Varma*, P. Wang*, X. Chen*, T. Chen*, S. Venugopalan, and Z. Wang, "Is Attention All That NeRF Needs?”, International Conference on Learning Representations (ICLR), 2023. [Paper] [Code]
  • D. Xu*, Y. Jiang*, P. Wang*, Z. Fan*, H. Shi, and Z. Wang, “SinNeRF: Training Neural Radiance Field on Complex Scenes from a Single Image”, European Conference on Computer Vision (ECCV), 2022. [Paper] [Code]

  • Prospective Students Shall Read More...

    Sponsor