Mansi Sakarvadia

Computer Science Ph.D Student

prof_pic.jpg

Hello! I am a Computer Science Ph.D. student at the University of Chicago, where I am co-advised by Ian Foster and Kyle Chard.

I develop machine learning interpretability methods. My research aims to systematically reverse engineer neural networks to interpret their weights. For example, much of my work focuses on localizing sources of model failure within weight-space and developing efficient methods to correct model behavior. My work is supported by a Department of Energy Computational Science Graduate Fellowship.

Prior to my Ph.D., I completed my Bachelors in Computer Science and Mathematics with a minor in Environmental Science at the University of North Carolina, Chapel Hill.

news

Sep 1, 2025 Had a great time at Lawrence Berekley National Laboratory’s ML and Analytics group this summer studying the limits of machine-learned operators for modeling PDEs.
Jul 16, 2025 Presented my poster “The False Promise of Super-Resolution of Machine-Learned Operators” at the CSGF Program Review in Washington, DC.
Apr 15, 2025 Was honored to have given a talk on my recent work on Mitigating Memorization in Language Models at the Midwest Speech and Language Days at University of Notre Dame!
Mar 1, 2025 Recent work on Mitigating Memorization in Language Models was accepted as a Top 5% Spotlight paper at International Conference on Learning Representations (ICLR) 2025! Check out a 5 min video summary of the work.
Jan 15, 2025 I was interviewed by the Department of Energy Science in Parallel podcast about the recent Nobel prizes in Physics and Chemistry and their implications for ML and the domain sciences.

selected publications

  1. Preprint
    The False Promise of Zero-Shot Super-Resolution in Machine-Learned Operators
    Mansi Sakarvadia, Kareem Hegazy, Amin Totounferoush, and 4 more authors
    2025
  2. Preprint
    Topology-Aware Knowledge Propagation in Decentralized Learning
    Mansi Sakarvadia, Nathaniel Hudson, Tian Li, and 2 more authors
    2025
  3. ICLR
    Mitigating Memorization In Language Models
    Mansi Sakarvadia, Aswathy Ajith, Arham Khan, and 6 more authors
    2025
    Spotlight (top 5%)
  4. Preprint
    SoK: On Finding Common Ground in Loss Landscapes Using Deep Model Merging Techniques
    Arham Khan, Todd Nief, Nathaniel Hudson, and 6 more authors
    2024
  5. BlackboxNLP
    Memory Injections: Correcting Multi-Hop Reasoning Failures during Inference in Transformer-Based Language Models
    Mansi Sakarvadia, Aswathy Ajith, Arham Khan, and 5 more authors
    2023
    Work accepted to BlackBoxNLP 2023.