EMNLP 2025

November 05, 2025

Suzhou, China

Would you like to see your presentation here, made available to a global audience of researchers?
Add your own presentation or have us affordably record your next conference.

Large language models (LLMs) display striking emergent behavior, notably solving arithmetic with only a few in-context examples (ICEs). Yet the computations that connect those examples to the answer remain opaque. We probe three open-weight LLMs, Pythia-12B, MPT-7B, and OPT-6.7B, on elementary arithmetic to illuminate how they process ICEs. Our study integrates activation patching, information-flow analysis, automatic circuit discovery, and the logit-lens perspective into a unified pipeline. Within this framework we isolate partial-sum representations in three-operand tasks, quantify their influence on final logits, and derive linear function vectors that characterize each operation and align with ICE-induced activations. Controlled ablations show that strict pattern consistency in the formatting of ICEs guides the models more strongly than the symbols chosen or even the factual correctness of the examples. By unifying four complementary interpretability tools, this work delivers one of the most comprehensive interpretability studies of LLM arithmetic to date, and the first on three-operand tasks. All code and datasets will be released.

Downloads

SlidesPaperTranscript English (automatic)

Next from EMNLP 2025

F²Bench: An Open-ended Fairness Evaluation Benchmark for LLMs with Factuality Considerations
poster

F²Bench: An Open-ended Fairness Evaluation Benchmark for LLMs with Factuality Considerations

EMNLP 2025

+3
Guanglai Gao and 5 other authors

05 November 2025

Stay up to date with the latest Underline news!

Select topic of interest (you can select more than one)

PRESENTATIONS

  • All Presentations
  • For Librarians
  • Resource Center
  • Free Trial
Underline Science, Inc.
1216 Broadway, 2nd Floor, New York, NY 10001, USA

© 2025 Underline - All rights reserved