
Manish Nagireddy
Ibm Research
text generation
fairness
knowledge graph
explainable ai
large language models
feature attribution
snlp: bias
transparency & privacy
snlp: language models
snlp: interpretability & analysis of nlp models
snlp: question answering
snlp: adversarial attacks & robustness
knowledge editing
input attribution
perturbation-based explanation
3
presentations
18
number of views
SHORT BIO
I am a Research Software Engineer at IBM Research and the MIT-IBM Watson AI Lab. My main research goal is to build trustworthy AI solutions. My research interests encompass several areas in machine learning and artificial intelligence from classical ML methods to natural language processing and the generative context. My current work focuses on use-case centered algorithmic auditing and evaluation, in the context of large language models. I am particularly interested in how the mapping of trustworthiness notions (e.g., fairness, explainability, robustness, etc.) translates to NLP settings.
Before joining IBM Research, I graduated from Carnegie Mellon University with a B.S. in statistics, machine learning, and computer science.
Presentations

Multi-Level Explanations for Generative Language Models
Lucas Monteiro Paes and 10 other authors

SocialStigmaQA: A Benchmark to Uncover Stigma Amplification in Generative Language Models | VIDEO
Manish Nagireddy and 3 other authors

Language Models in Dialogue: Conversational Maxims for Human-AI Interactions
Erik Miehling and 5 other authors