Mansi Sakarvadia

Computer Science Ph.D Student

prof_pic.jpg

Hello! I am a Computer Science Ph.D. student at the University of Chicago, where I am co-advised by Ian Foster and Kyle Chard.

I research and develop adaptive, scalable, and fault-tolerant methods for machine learning (ML). My research is motivated by the challenges of modern ML lifecycles, where models are being developed for and deployed in increasingly complex and dynamic computational ecosystems. These fast-evolving modeling landscapes require effective design, measurement, and management.

In the short-term, my work focuses on studying, preventing, and efficiently correcting failure modes within ML lifecycles. For example, I have developed methods to enable fast adaption of LMs to mitigate unwanted behavior, fault-tolerant training over decentralized data, and scalable scientific modeling. In the long-term, my goal is to enable resilient machine learning.

My work is supported by a Department of Energy Computational Science Graduate Fellowship. Prior to my Ph.D., I completed my Bachelors in Computer Science and Mathematics at UNC, Chapel Hill.

news

Sep 1, 2025 Had a great time at Lawrence Berekley National Laboratory’s ML and Analytics group this summer studying the limits of machine-learned operators for modeling PDEs.
Jul 16, 2025 Presented my poster “The False Promise of Super-Resolution of Machine-Learned Operators” at the CSGF Program Review in Washington, DC.
Apr 15, 2025 Was honored to have given a talk on my recent work on Mitigating Memorization in Language Models at the Midwest Speech and Language Days at University of Notre Dame!
Mar 1, 2025 Recent work on Mitigating Memorization in Language Models was accepted as a Top 5% Spotlight paper at International Conference on Learning Representations (ICLR) 2025! Check out a 5 min video summary of the work.
Jan 15, 2025 I was interviewed by the Department of Energy Science in Parallel podcast about the recent Nobel prizes in Physics and Chemistry and their implications for ML and the domain sciences.

selected publications

  1. Preprint
    The False Promise of Zero-Shot Super-Resolution in Machine-Learned Operators
    Mansi Sakarvadia, Kareem Hegazy, Amin Totounferoush, and 4 more authors
    2025
  2. Preprint
    Topology-Aware Knowledge Propagation in Decentralized Learning
    Mansi Sakarvadia, Nathaniel Hudson, Tian Li, and 2 more authors
    2025
  3. ICLR
    Mitigating Memorization In Language Models
    Mansi Sakarvadia, Aswathy Ajith, Arham Khan, and 6 more authors
    2025
    Spotlight (top 5%)
  4. Preprint
    SoK: On Finding Common Ground in Loss Landscapes Using Deep Model Merging Techniques
    Arham Khan, Todd Nief, Nathaniel Hudson, and 6 more authors
    2024
  5. BlackboxNLP
    Memory Injections: Correcting Multi-Hop Reasoning Failures during Inference in Transformer-Based Language Models
    Mansi Sakarvadia, Aswathy Ajith, Arham Khan, and 5 more authors
    2023
    Work accepted to BlackBoxNLP 2023.