profile picture

Nanyun Peng

Associate Professor @ UCLA

event extraction

large language models

dataset

evaluation

information extraction

question answering

paraphrasing

story generation

commonsense

natural language processing

benchmark

data augmentation

natural language generation

factuality

event argument extraction

66

presentations

55

number of views

1

citations

SHORT BIO

Nanyun is an Assistant Professor at the Computer Science Department, University of California, Los Angeles. Her research goals aim to build robust and generalizable Natural Language Processing (NLP) tools that lower the communication barriers and enable AI agents to become companions for humans. With these goals in mind, she has been focusing on several research topics, including creative language generation, low-resource information extraction, and zero-shot cross-lingual transfer. She got her PhD in Computer Science at Johns Hopkins University, Center for Language and Speech Processing, after that, she spent three awesome years at University of Southern California as a Research Assistant Professor at the Computer Science Department, and a Research Lead at the Information Sciences Institute. She has backgrounds in Linguistics and Economics and held BAs in both.

Presentations

Measuring Psychological Depth in Language Models

Fabrice Harel-Canada and 6 other authors

Evaluating LLMs' Capability in Satisfying Lexical Constraints

Bingxuan Li and 4 other authors

SPEED++: A Multilingual Event Extraction Framework for Epidemic Prediction and Preparedness

Tanmay Parekh and 8 other authors

VDebugger: Harnessing Execution Feedback for Debugging Visual Programs

Xueqing Wu and 6 other authors

Explaining and Improving Contrastive Decoding by Extrapolating the Probabilities of a Huge and Hypothetical LM

Haw-Shiuan Chang and 4 other authors

QUDSELECT: Selective Decoding for Questions Under Discussion Parsing

Ashima Suvarna and 4 other authors

Re-ReST: Reflection-Reinforced Self-Training for Language Agents

Zi-Yi Dou and 4 other authors

Model Editing Harms General Abilities of Large Language Models: Regularization to the Rescue

Jia-Chen Gu and 6 other authors

Synchronous Faithfulness Monitoring for Trustworthy Retrieval-Augmented Generation

Di Wu and 4 other authors

LLM Self-Correction with DeCRIM: Decompose, Critique, and Refine for Enhanced Following of Instructions with Multiple Constraints

Thomas Palmeira Ferraz and 9 other authors

LLM-A*: Large Language Model Enhanced Incremental Heuristic Search on Path Planning

Silin Meng and 4 other authors

Tracking the Newsworthiness of Public Documents

Alexander Spangher and 5 other authors

PhonologyBench: Evaluating Phonological Skills of Large Language Models

Ashima Suvarna and 2 other authors

Argument-Aware Approach To Event Linking

I-Hung Hsu and 6 other authors

CaLM: Contrasting Large and Small Language Models to Verify Grounded Generation

I-Hung Hsu and 6 other authors

Improving Event Definition Following For Zero-Shot Event Detection

Zefan Cai and 8 other authors

Stay up to date with the latest Underline news!

Select topic of interest (you can select more than one)

PRESENTATIONS

  • All Presentations
  • For Librarians
  • Resource Center
  • Free Trial
Underline Science, Inc.
1216 Broadway, 2nd Floor, New York, NY 10001, USA

© 2025 Underline - All rights reserved