profile picture

Arman Cohan

large language models

summarization

retrieval

evaluation

data augmentation

question answering

text generation

instruction following

information retrieval

readability

natural language processing

benchmarking

generation

multi-document summarization

multilingual

16

presentations

3

number of views

SHORT BIO

Arman Cohan is an Assistant Professor of Computer Science at Yale University and a faculty Research Scientist at the Allen Institute for AI (AI2). His research spans various problems at the intersection of Natural Language Processing and Machine Learning, including Language Modeling, Representation Learning, Generation, and their applications to specialized domains include science. His research has been recognized with multiple awards, including a best paper award at EMNLP, an outstanding paper award at EACL, and an honorable mention at COLING. Prior to joining Yale, he was a Research Scientist at the Allen Institute for AI (AI2) and an Affiliate Assistant Professor at University of Washington.

Presentations

TAIL: A Toolkit for Automatic and Realistic Long-Context Large Language Model Evaluation

Gefei Gu and 4 other authors

Bayesian Calibration of Win Rate Estimation with LLM Evaluators

Yicheng Gao and 3 other authors

Calibrating Long-form Generations From Large Language Models

Yukun Huang and 4 other authors

P-FOLIO: Evaluating and Improving Logical Reasoning with Abundant Human-Written Reasoning Chains

Simeng Han and 15 other authors

SciDQA: A Deep Reading Comprehension Dataset over Scientific Papers

Shruti Singh and 2 other authors

TaPERA: Enhancing Faithfulness and Interpretability in Long-Form Table QA by Content Planning and Execution-based Reasoning

Yilun Zhao and 3 other authors

Quantifying Contamination in Evaluating Code Generation Capabilities of Language Models

Martin Riddell and 2 other authors

Benchmarking Generation and Evaluation Capabilities of Large Language Models for Instruction Controllable Summarization

Yixin Liu and 9 other authors

Investigating Data Contamination in Modern Benchmarks for Large Language Models

Chunyuan Deng and 4 other authors

On Learning to Summarize with Large Language Models as References

Yixin Liu and 7 other authors

On the Benefits of Fine-Grained Loss Truncation: A Case Study on Factuality in Summarization

Lorenzo Jaime Flores and 1 other author

When do Generative Query and Document Expansions Fail? A Comprehensive Study Across Methods, Retrievers, and Datasets

Orion Weller and 6 other authors

TESS: Text-to-Text Self-Conditioned Simplex Diffusion

Rabeeh Karimi mahabadi and 6 other authors

Investigating Table-to-Text Generation Capabilities of Large Language Models in Real-World Information Seeking Scenarios

Yilun Zhao and 5 other authors

Medical Text Simplification: Optimizing for Readability with Unlikelihood Training and Reranked Beam Search Decoding

Lorenzo Jaime Flores and 4 other authors

Evaluating the Impact of Retrieval on Multi-document Summarization

Arman Cohan

Stay up to date with the latest Underline news!

Select topic of interest (you can select more than one)

PRESENTATIONS

  • All Lectures
  • For Librarians
  • Resource Center
  • Free Trial
Underline Science, Inc.
1216 Broadway, 2nd Floor, New York, NY 10001, USA

© 2025 Underline - All rights reserved