Hima Lakkaraju

Contact
hlakkaraju@hbs.eduhlakkaraju@seas.harvard.edu
442 Morgan Hall
337 Maxwell Dworkin
Wikipedia
@hima_lakkaraju
lvhimabindu
I am an Assistant Professor at Harvard University with appointments in the Business School and the Department of Computer Science.
My research interests lie within the broad area of trustworthy machine learning. More specifically, I focus on improving the interpretability, fairness, robustness, and reasoning capabilities of different kinds of ML models including language models and other large pre-trained models. My research addresses the following fundamental questions pertaining to human and algorithmic decision-making:
- How to build fair, robust, and interpretable models that can aid human decision-making?
- How to ensure that models and their explanations are robust to adversarial attacks?
- How to train and evaluate models in the presence of missing counterfactuals?
- How to detect and correct underlying biases in human decisions and algorithmic predictions?
These questions have far-reaching implications in domains involving high-stakes decisions such as health care, policy, law, and business.
I lead the AI4LIFE research group at Harvard and I recently co-founded the Trustworthy ML Initiative (TrustML) to help lower entry barriers into trustworthy ML and bring together researchers and practitioners working in the field.
My research is being generously supported by NSF, Google, Amazon, JP Morgan, Bayer, Harvard Data Science Initiative, and D^3 Insitute at Harvard. Prior to my stint at Harvard, I received my PhD in Computer Science from Stanford University.
For more details about me and my research, please check out my CV.
NOTE: I am looking for motivated graduate and undergraduate students and postdocs who are broadly interested in trustworthy machine learning and large pre-trained models. If you are excited about this line of research and would like to work with me, please read this before contacting me.
Selected Achievements
- NSF CAREER Award, 2023
- Named Kavli Fellow 2023 by National Academy of Sciences
- My short course on Explainable AI (hosted by Stanford University) is now available on Youtube, 2022
- Outstanding Paper Award Honorable Mention, NeurIPS Workshop on Trustworthy and Socially Responsible Machine Learning, 2022
- JP Morgan Faculty Research Award, 2022
- Best Paper Award, ICML Workshop on Interpretable ML in Healthcare, 2022
- Released the first version of OpenXAI, a light-weight open source library to evaluate and benchmark post hoc explanation methods, 2022
- Amazon Research Award, 2021
- Google AI for Social Good Research Award, 2021
- Best Paper Runner Up, ICML Workshop on Algorithmic Recourse, 2021
- Google Research Award, 2020
- Hoopes prize for undergraduate thesis mentoring, Harvard University, 2020
- Co-founded Trustworthy ML Initiative to enable easy access to resources on trustworthy ML & to build a community of researchers/practitioners, 2020
- Named one of the world's 35 innovators under 35 by MIT Tech Review, 2019
- Named one of the world's top innovators to watch by Vanity Fair, 2019
- Selected for the prestigious Cowles fellowship by Yale University, 2018
- INFORMS Best Data Mining Paper Award, 2017
- Microsoft Research Dissertation Grant, 2017
- Named a Rising Star in Computer Science, 2016
- Google Anita Borg Fellowship , 2015
- Stanford Graduate Fellowship, 2013-17
- Eminence and Excellence Award, IBM Research, 2012
- Research Division Award, IBM Research, 2012
- Best Paper Award, SIAM International Conference on Data Mining, 2011
Upcoming and Recent Talks
-
07.2023 ICML Workshop on Interpretable ML in Healthcare -
07.2023 ICML Workshop on Counterfactuals in Minds and Machines -
05.2023 ICLR Workshop on Trustworthy and Reliable Large-Scale Machine Learning Models -
05.2023 Responsible AI Workshop at Carnegie Mellon University -
04.2023 Guest Lecture at Carnegie Mellon University -
04.2023 Mind and Machine Intelligence Summit, UC Santa Barbara -
04.2023 ACM India Bootcamp on Responsible Computing -
03.2023 Cornell University and Weill Cornell Medicine -
03.2023 Guest Lecture at UC Berkeley -
03.2023 Kavli Frontiers of Science Symposium -
03.2023 Cohere AI -
02.2023 Keynote at AAAI Workshop on Representation Learning for Responsible Human-Centric AI -
02.2023 AAAI Workshop on Deployable AI -
12.2022 Berkman Klein Center, Harvard University -
11.2022 Keynote at Women in Machine Learning (WiML) Workshop Co-located with NeurIPS, 2022 -
11.2022 Machine Learning for Health (ML4H) Workshop Co-located with NeurIPS, 2022 -
11.2022 Simons Institute (Berkeley) Workshop on Societal Considerations and Applications -
11.2022 MIT Initiative on the Digital Economy (IDE) Seminar Series -
11.2022 Harvard Data Science Initiative's Annual Conference -
10.2022 ECCV Workshop on Adversarial Robustness in the Real World -
08.2022 Stanford Center for AI Safety -
08.2022 Amazon Alexa Rising Star Speaker Series -
06.2022 CVPR Workshop on Explainable AI for Computer Vision -
05.2022 ICLR Workshop on Privacy, Accountability, Interpretability, Robustness, Reasoning on Structured Data -
05.2022 Keynote at WWW Workshop on Explainable AI in Health -
05.2022 Fiddler AI Fireside Chat -
04.2022 AI and the Economy (U.S. Department of Commerce, National Institute of Standards and Technology, Stanford HAI, and the FinRegLab) -
04.2022 Stanford Human-Centered Artificial Intelligence (HAI) Conference -
04.2022 Stanford Digital Econ Seminar -
03.2022 University of Southern California
Research
Publications
- See Google Scholar Page for latest preprints
- * below indicates equal contribution
-
TalktoModel: Explaining Machine Learning Models with Interactive Natural Language Conversations
Dylan Slack, Satyapriya Krishna, Himabindu Lakkaraju*, Sameer Singh*
Nature Machine Intelligence, 2023.
Outstanding Paper Award Honorable Mention, NeurIPS Workshop on Trustworthy and Socially Responsible Machine Learning, 2022.
pdf -
Evaluating Explainability for Graph Neural Networks
Chirag Agarwal, Owen Queen, Himabindu Lakkaraju, Marinka Zitnik
Nature Scientific Data, 2023.
pdf -
When Does Uncertainty Matter?: Understanding the Impact of Predictive Uncertainty in ML Assisted Decision Making
Sean McGrath, Parth Mehta, Alexandra Zytek, Isaac Lage, Himabindu Lakkaraju
Transactions on Machine Learning Research (TMLR), 2023.
pdf -
Towards Bridging the Gaps between the Right to Explanation and the Right to be Forgotten
Satyapriya Krishna*, Jiaqi Ma*, Himabindu Lakkaraju
International Conference on Machine Learning (ICML), 2023
pdf -
On the Impact of Actionable Explanations on Social Segregation
Ruijiang Gao, Himabindu Lakkaraju
International Conference on Machine Learning (ICML), 2023
pdf -
On Minimizing the Impact of Dataset Shifts on Actionable Explanations
Anna Meyer*, Dan Ley*, Suraj Srinivas, Himabindu Lakkaraju
Conference on Uncertainty in Artificial Intelligence (UAI), 2023
pdf -
Probabilistically Robust Recourse: Navigating the Trade-offs between Costs and Robustness in Algorithmic Recourse
Martin Pawelczyk, Teresa Datta, Johannes van-den-Heuvel, Gjergji Kasneci, Himabindu Lakkaraju
International Conference on Learning Representations (ICLR), 2023
pdf -
On the Privacy Risks of Algorithmic Recourse
Martin Pawelczyk, Himabindu Lakkaraju, Seth Neel
International Conference on Artificial Intelligence and Statistics (AISTATS), 2023
pdf -
Which Explanation Should I Choose? A Function Approximation Perspective to Characterizing Post hoc Explanations
Tessa Han, Suraj Srinivas, Himabindu Lakkaraju
Advances in Neural Information Processing Systems (NeurIPS), 2022.
Best Paper Award, ICML Workshop on Interpretable Machine Learning in Healthcare, 2022.
pdf -
Flatten the Curve: Efficiently Training Low-Curvature Neural Networks
Suraj Srinivas, Kyle Matoba, Himabindu Lakkaraju, Francois Fleuret
Advances in Neural Information Processing Systems (NeurIPS), 2022.
pdf -
OpenXAI: Towards a Transparent Evaluation of Model Explanations
Chirag Agarwal, Satyapriya Krishna, Eshika Saxena, Martin Pawelczyk, Nari Johnson, Isha Puri, Marinka Zitnik, Himabindu Lakkaraju
Advances in Neural Information Processing Systems (NeurIPS), 2022.
pdf -
Data Poisoning Attacks on Off-Policy Evaluation Methods
Elita Lobo, Harvineet Singh, Marek Petrik, Cynthia Rudin, Himabindu Lakkaraju
Conference on Uncertainty in Artificial Intelligence (UAI), 2022.
Oral Presentation [Top 5%]
pdf -
Exploring Counterfactual Explanations Through the Lens of Adversarial Examples: A Theoretical and Empirical Analysis.
Martin Pawelczyk, Chirag Agarwal, Shalmali Joshi, Sohini Upadhyay, Himabindu Lakkaraju
International Conference on Artificial Intelligence and Statistics (AISTATS), 2022.
pdf -
Probing GNN Explainers: A Rigorous Theoretical and Empirical Analysis of GNN Explanation Methods.
Chirag Agarwal, Marinka Zitnik, Himabindu Lakkaraju
International Conference on Artificial Intelligence and Statistics (AISTATS), 2022.
pdf -
Fairness via Explanation Quality: Evaluating Disparities in the Quality of Post hoc Explanations.
Jessica Dai, Sohini Upadhyay, Ulrich Aivodji, Stephen Bach, Himabindu Lakkaraju
AAAI/ACM Conference on AI, Society, and Ethics (AIES), 2022.
pdf -
Towards Robust Off-Policy Evaluation via Human Inputs.
Harvineet Singh, Shalmali Joshi, Finale Doshi-Velez, Himabindu Lakkaraju
AAAI/ACM Conference on AI, Society, and Ethics (AIES), 2022.
pdf -
A Human-Centric Take on Model Monitoring.
Murtuza N Shergadwala, Himabindu Lakkaraju, Krishnaram Kenthapadi
AAAI Conference on Human Computation and Crowdsourcing (HCOMP), 2022.
pdf -
Towards the Unification and Robustness of Post hoc Explanation Methods.
Sushant Agarwal, Shahin Jabbari, Chirag Agarwal*, Sohini Upadhyay*, Steven Wu, Himabindu Lakkaraju
Symposium on Foundations of Responsible Computing (FORC), 2022.
pdf -
Towards Robust and Reliable Algorithmic Recourse.
Sohini Upadhyay*, Shalmali Joshi*, Himabindu Lakkaraju
Advances in Neural Information Processing Systems (NeurIPS), 2021.
Best Paper Runner Up, ICML Workshop on Algorithmic Recourse, 2021.
pdf -
Reliable Post hoc Explanations: Modeling Uncertainty in Explainability.
Dylan Slack, Sophie Hilgard, Sameer Singh, Himabindu Lakkaraju
Advances in Neural Information Processing Systems (NeurIPS), 2021.
pdf -
Counterfactual Explanations Can Be Manipulated
Dylan Slack, Sophie Hilgard, Himabindu Lakkaraju, Sameer Singh
Advances in Neural Information Processing Systems (NeurIPS), 2021.
pdf -
Learning Models for Algorithmic Recourse
Alexis Ross, Himabindu Lakkaraju, Osbert Bastani
Advances in Neural Information Processing Systems (NeurIPS), 2021.
pdf -
Towards the Unification and Robustness of Perturbation and Gradient Based Explanations.
Sushant Agarwal, Shahin Jabbari, Chirag Agarwal*, Sohini Upadhyay*, Steven Wu, Himabindu Lakkaraju
International Conference on Machine Learning (ICML), 2021.
pdf -
Towards a Unified Framework for Fair and Stable Graph Representation Learning.
Chirag Agarwal, Himabindu Lakkaraju, Marinka Zitnik
Conference on Uncertainty in Artificial Intelligence (UAI), 2021.
pdf -
Fair influence maximization: A welfare optimization approach.
Aida Rahmattalabi, Shahin Jabbari, Himabindu Lakkaraju, Phebe Vayanos, Eric Rice, Milind Tambe
AAAI International Conference on Artificial Intelligence (AAAI), 2021.
pdf -
Does Fair Ranking Improve Minority Outcomes? Understanding the Interplay of Human and Algorithmic Biases in Online Hiring.
Tom Suhr, Sophie Hilgard, Himabindu Lakkaraju
AAAI/ACM Conference on Artificial Intelligence, Ethics, and Society (AIES), 2021.
pdf -
Beyond Individualized Recourse: Interpretable and Interactive Summaries of Actionable Recourses
Kaivalya Rawal and Himabindu Lakkaraju.
Advances in Neural Information Processing Systems (NeurIPS), 2020.
pdf -
Incorporating Interpretable Output Constraints in Bayesian Neural Networks
Wanqian Yang, Lars Lorch, Moritz Gaule, Himabindu Lakkaraju, Finale Doshi-Velez.
Advances in Neural Information Processing Systems (NeurIPS), 2020.
Spotlight Presentation [Top 3%]
pdf
-
Robust and Stable Black Box Explanations.
Himabindu Lakkaraju, Nino Arsov, Osbert Bastani.
International Conference on Machine Learning (ICML), 2020.
pdf -
Fooling LIME and SHAP: Adversarial Attacks on Post hoc Explanation Methods.
Dylan Slack, Sophie Hilgard, Emily Jia, Sameer Singh, Himabindu Lakkaraju.
AAAI/ACM Conference on Artificial Intelligence, Ethics, and Society (AIES), 2020.
Oral Presentation
pdf
Press: deeplearning.ai | Harvard Business Review -
"How do I fool you?": Manipulating User Trust via Misleading Black Box Explanations.
Himabindu Lakkaraju, Osbert Bastani.
AAAI/ACM Conference on Artificial Intelligence, Ethics, and Society (AIES), 2020.
Oral Presentation
pdf -
Faithful and Customizable Explanations of Black Box Models.
Himabindu Lakkaraju, Ece Kamar, Rich Carauna, Jure Leskovec.
AAAI/ACM Conference on Artificial Intelligence, Ethics, and Society (AIES), 2019.
Oral Presentation
pdf Human Decisions and Machine Predictions.
Jon Kleinberg, Himabindu Lakkaraju, Jure Leskovec, Jens Ludwig, Sendhil Mullainathan.
Quarterly Journal of Economics (QJE), 2018.
Featured in MIT Technology Review, Harvard Business Review, The New York Times,
and as Research Spotlight on National Bureau of Economics front page .
pdfThe Selective Labels Problem: Evaluating Algorithmic Predictions in the Presence of Unobservables.
Himabindu Lakkaraju, Jon Kleinberg, Jure Leskovec, Jens Ludwig, Sendhil Mullainathan.
ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD), 2017.
Oral Presentation
pdfLearning Cost-Effective and Interpretable Treatment Regimes.
Himabindu Lakkaraju, Cynthia Rudin.
International Conference on Artificial Intelligence and Statistics (AISTATS), 2017.
INFORMS Data Mining Best Paper Award .
Invited Talk at INFORMS Annual Meeting .
pdfIdentifying Unknown Unknowns in the Open World: Representations and Policies for Guided Exploration.
Himabindu Lakkaraju, Ece Kamar, Rich Caruana, Eric Horvitz.
AAAI Conference on Artificial Intelligence (AAAI), 2017.
Featured in Bloomberg Technology .
pdfInterpretable and Explorable Approximations of Black Box Models.
Himabindu Lakkaraju, Ece Kamar, Rich Caruana, Jure Leskovec.
KDD Workshop on Fairness, Accountability, and Transparency in Machine Learning (FAT ML), 2017.
Invited Talk at INFORMS Annual Meeting .
pdfConfusions over Time: An Interpretable Bayesian Model to Characterize Trends in Decision Making.
Himabindu Lakkaraju, Jure Leskovec.
Advances in Neural Information Processing Systems (NIPS), 2016.
pdfInterpretable Decision Sets: A Joint Framework for Description and Prediction.
Himabindu Lakkaraju, Stephen H. Bach, Jure Leskovec.
ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD), 2016.
Invited Talk at INFORMS Annual Meeting .
pdfMining Big Data to Extract Patterns and Predict Real-Life Outcomes.
Michal Kosinki, Yilun Wang, Himabindu Lakkaraju, Jure Leskovec.
Psychological Methods, 2016.
pdfLearning Cost-Effective and Interpretable Regimes for Treatment Recommendation.
Himabindu Lakkaraju, Cynthia Rudin.
NIPS Workshop on Interpretable Machine Learning in Complex Systems, 2016.
pdfLearning Cost-Effective and Interpretable Treatment Regimes for Judicial Bail Decisions.
Himabindu Lakkaraju, Cynthia Rudin.
NIPS Symposium on Machine Learning and the Law, 2016.
pdfDiscovering Unknown Unknowns of Predictive Models.
Himabindu Lakkaraju, Ece Kamar, Rich Caruana, Eric Horvitz.
NIPS Workshop on Reliable Machine Learning in the Wild, 2016.
pdfA Machine Learning Framework to Identify Students at Risk of Adverse Academic Outcomes.
Himabindu Lakkaraju, Everaldo Aguiar, Carl Shan, David Miller, Nasir Bhanpuri, Rayid Ghani.
ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD), 2015.
Oral Presentation
pdfA Bayesian Framework for Modeling Human Evaluations.
Himabindu Lakkaraju, Jure Leskovec, Jon Kleinberg, Sendhil Mullainathan.
SIAM International Conference on Data Mining (SDM) , 2015.
Oral Presentation
pdfWho, When, and Why: A Machine Learning Approach to Prioritizing Students at Risk of not Graduating High School on Time.
Everaldo Aguiar, Himabindu Lakkaraju, Nasir Bhanpuri, David Miller, Ben Yuhas, Kecia Addison, Rayid Ghani.
Learning Analytics and Knowledge Conference (LAK), 2015.
pdfWhat's in a name ? Understanding the Interplay Between Titles, Content, and Communities in Social Media.
Himabindu Lakkaraju, Julian McAuley, Jure Leskovec.
International AAAI Conference on Weblogs and Social Media (ICWSM), 2013.
Oral Presentation
Featured in Time, Forbes, Phys.Org, Business Insider .
pdfDynamic Multi-Relational Chinese Restaurant Process for Analyzing Influences on Users in Social Media.
Himabindu Lakkaraju, Indrajit Bhattacharya, Chiranjib Bhattacharyya.
IEEE International Conference on Data Mining (ICDM), 2012.
Oral Presentation
pdfTEM: a novel perspective to modeling content on microblogs.
Himabindu Lakkaraju, Hyung-Il Ahn.
International World Wide Web Conference (WWW), short paper, 2012.
pdfExploiting Coherence for the Simultaneous Discovery of Latent Facets and associated Sentiments.
Himabindu Lakkaraju, Chiranjib Bhattacharyya, Indrajit Bhattacharya, Srujana Merugu.
SIAM International Conference on Data Mining (SDM), 2011.
Best Paper Award.
pdfAttention prediction on social media brand pages.
Himabindu Lakkaraju, Jitendra Ajmera.
ACM Conference on Information and Knowledge Management (CIKM), 2011.
pdfSmart news feeds for social networks using scalable joint latent factor models.
Himabindu Lakkaraju, Angshu Rai, Srujana Merugu.
International World Wide Web Conference (WWW), short paper, 2011.
pdf
Patents
Extraction and Grouping of Feature Words.
Himabindu Lakkaraju, Chiranjib Bhattacharyya, Sunil Aravindam, Kaushik Nath.
US8484228
Enhancing knowledge bases using rich social media.
Jitendra Ajmera, Shantanu Ravindra Godbole, Himabindu Lakkaraju, Bernard Andrew Roden, Ashish Verma.
US20130224714
Advising
I am very fortunate to be working with the following core group of students, interns, postdocs, and research affiliates
- Suraj Srinivas (Postdoc, Harvard University)
- Jiaqi Ma (Postdoc, Harvard University)
- Tessa Han (PhD Student, Harvard University)
- Satyapriya Krishna (PhD Student, Harvard University)
- Usha Bhalla (PhD Student, Harvard University)
- Dan Ley (PhD Student, Harvard University)
- Paul Hamilton (PhD Student, Harvard University)
- Alex Oesterling (PhD Student, Harvard University); Co-advised with Flavio Calmon
- Dylan Slack (PhD Student, UC Irvine); Co-advised with Sameer Singh
- Eshika Saxena (Undergrad, Harvard University)
- Chirag Agarwal (Research Scientist, Adobe Research; Research Affiliate, Harvard University)
- Martin Pawelczyk (PhD Student, University of Tubingen; Research Fellow, Harvard University)
- Davor Ljubenkov (Fullbright Scholar; Research Fellow, Harvard University)
Alumni (Past Advisees, Close Collaborators, and Visitors):
- Chirag Agarwal (Postdoc, Harvard University --> Research Scientist, Adobe Research)
- Alexis Ross (Undergraduate Student, Harvard University -- Winner of Hoopes Prize for Best Undergrad Thesis --> MIT EECS PhD)
- Isha Puri (Undergraduate Student, Harvard University --> MIT EECS PhD)
- Jessica Dai (Undergraduate Student, Brown University --> UC Berkeley EECS PhD)
- Aditya Karan (Masters Student, Harvard University --> PhD Student, UIUC CS)
- Kaivalya Rawal (Masters Student, Harvard University --> Fiddler AI)
- Ethan Kim (Undergraduate Student, Harvard University --> Cyndx)
- Sophie Hilgard (PhD Student, Harvard University --> Research Scientist, Twitter)
- Sushant Agarwal (Masters Student, University of Waterloo --> PhD Student, Northeastern University)
- Harvineet Singh (PhD Student, New York University; Research Intern, Harvard University) --> Postdoc UCSF/UC Berkeley
- Elita Lobo (PhD Student, UMass Amherst; Research Intern, Harvard University)
- Anna Meyer (PhD Student, University of Wisconsin; Research Intern, Harvard University)
- Ruijiang Gao (PhD Student, University of Texas at Austin; Research Intern, Harvard University)
- Vishwali Mhasawade (PhD Student, New York University; Research Intern, Harvard University)
- Tom Suhr (MS Student, TU Berlin; Research Fellow, Harvard University) --> PhD Student, Max Planck Institute
Teaching
Topics in Machine Learning: Interpretability and Explainability
Instructor
Harvard University, 2023.
Introduction to Technology and Operations Management
Instructor
Harvard University, 2022.
Topics in Machine Learning: Interpretability and Explainability
Instructor
Harvard University, 2021.
Introduction to Technology and Operations Management
Instructor
Harvard University, 2020.
Introduction to Machine Learning for Social Scientists
Instructor
Harvard University, 2020.
Topics in Machine Learning: Interpretability and Explainability
Instructor
Harvard University, 2019.
Introduction to Data Science
Guest Lecture
Stanford Law School, 2016.
Probability with Mathemagic
Co-Instructor
Stanford Splash Initiative for High School Students, 2016.
Mining Massive Datasets Course
Teaching Assistant
Stanford Computer Science, 2016.
Submodular Optimization
Guest Lecture
Mining Massive Datasets Course, Stanford, 2016.
Introduction to Python Programming
Co-Instructor
Stanford Girls Teaching Girls to Code Initiative for High School Students, 2015.
Mathematics and Science
Tutor
Dreamcatchers Non-Profit Organization, Palo Alto, 2015.
Social and Information Network Analysis Course
Head Teaching Assistant
Stanford Computer Science, 2014.
Machine Learning Course
Teaching Assistant
Indian Institute of Science, 2010.
English and Mathematics
Tutor
UNICEF's Teach India Initiative, 2008 - 2010.
Object Oriented Programming
Co-Instructor
Visvesvaraya Technological University, 2007.
Introduction to Databases
Co-Instructor
Visvesvaraya Technological University, 2007.