About Me
Hello! I am a PhD student in MIT LIDS, working on trustworthy machine learning with interests in high-stakes domains like healthcare. I'm lucky to be co-advised by Professors Marzyeh Ghassemi and Collin Stultz. In 2023, I graduated from Princeton University with a B.S.E. in Operations Research and Financial Engineering, along with minors in Cognitive Science, Computer Science, Linguistics, and Statistics & Machine Learning. My senior thesis, supervised by the incredible Professor Christiane Fellbaum, explored NLP techniques for detecting and editing stigmatizing language in medical records.
My research focuses on reliable, responsible, and trustworthy Machine Learning, and my work spans group robustness, GenAI Agents, chain-of-thought, and backdoor attacks. I use tools and frameworks from Optimization, Probability, and Statistics. My work has been published at venues such as ACM FAccT and IEEE.
This past summer, I interned at IBM Research, where I investigated inverted model outputs to identify gaps in reasoning chains and improve hallucination detection for large language models.
You can reach me at abinitha@mit.edu. I'd love to hear from you!
News
- Feb 2026: Our paper LEIA is released on arXiv!
- Oct 2025: Presented an ML Tea Talk on trace inversion.
- Sept 2025: Excited to serve as Program Chair of the 2026 LIDS Student Conference!
- Aug 2025: Completed my summer internship at IBM Research!
- June 2025: Our work is covered by MIT News and other press.
- June 2025: Presented at FAccT '25!
- June 2025: Paper released on arXiv!
- May 2025: Excited to join the AIES PC!
- May 2025: MedPerturb project website online!
- April 2025: Presented @ the MIT EECS Town Hall for SERC.
- April 2025: Paper accepted to FAccT '25!
- April 2025: Presented @ the MIT IMES Seminar Series.
- Aug 2024: Started my PhD at MIT!
Research Highlights
The MedPerturb Dataset: What Non-Content Perturbations Reveal About Human and Clinical LLM Decision Making
The Medium is the Message: How Non-Clinical Information Shapes Clinical Decisions in LLMs
PanDa Game: Optimized Privacy-Preserving Publishing of Individual-Level Pandemic Data Based on a Game Theoretic Model
What seems to be the problem? Stigmatizing language in patient medical notes
Papers
- Abinitha Gourabathina, Hyewon Jeong, Teya Bergamaschi, Marzyeh Ghassemi, and Collin Stultz, 2026. Robustness Beyond Known Groups with Low-rank Adaptation. arXiv preprint arXiv:2602.06924. (Preprint).
[Paper] [Code] - Abinitha Gourabathina, Yuexing Hao, Walter Gerych, and Marzyeh Ghassemi, 2025. The MedPerturb Dataset: What Non-Content Perturbations Reveal About Human and Clinical LLM Decision Making. arXiv preprint arXiv:2506.17163. (Preprint).
[Paper] [Website] [Code] - Abinitha Gourabathina, Walter Gerych, Eileen Pan, and Marzyeh Ghassemi. 2025. The Medium is the Message: How Non-Clinical Information Shapes Clinical Decisions in LLMs. In Proceedings of the 2025 ACM Conference on Fairness, Accountability, and Transparency (FAccT '25).
[Paper] [Code] - Abinitha Gourabathina, Zhiyu Wan, J. Thomas Brown, Chao Yan, Bradley A. Malin, 2023. PanDa Game: Optimized Privacy-Preserving Publishing of Individual-Level Pandemic Data Based on a Game Theoretic Model. In IEEE Transactions on NanoBioscience, vol. 22, no. 4, pp. 808-817, Oct. 2023.
[Paper] [Code] - Abinitha Gourabathina, Zhiyu Wan, J. Thomas Brown, Chao Yan, Bradley A. Malin, 2022. PanDa Game: Optimized Privacy-Preserving Publishing of Individual-Level Pandemic Data Based on a Game Theoretic Model. In Proceedings of the 2022 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Las Vegas, NV, USA, 2022, pp. 961-968.
[Paper] - Abinitha Gourabathina (Senior thesis, Princeton University). What seems to be the problem? Stigmatizing language in patient medical notes. Princeton DataSpace. http://arks.princeton.edu/ark:/88435/dsp01cv43p110t
[Paper]