profile picture

Yohei Oseki

Associate Professor @ University of Tokyo, Department of Language and Information Sciences, Tokyo, Japan

cognitive modeling

syntactic supervision

language model

syntax

psycholinguistics

computational psycholinguistics

language models

composition

efficiency

inductive bias

rnng

large language model

neural language model

targeted syntactic evaluation

targeted syntactic evaluations

11

presentations

18

number of views

Presentations

Can Language Models Induce Grammatical Knowledge from Indirect Evidence?

Miyu Oba and 6 other authors

Tree-Planted Transformers: Unidirectional Transformer Language Models with Implicit Syntactic Supervision

Ryo Yoshida and 2 other authors

Modeling Overregularization in Children with Small Language Models

Akari Haga and 6 other authors

Emergent Word Order Universals from Cognitively-Motivated Language Models

Tatsuki Kuribayashi and 5 other authors

Psychometric Predictive Power of Large Language Models

Tatsuki Kuribayashi and 2 other authors

JBLiMP: Japanese Benchmark of Linguistic Minimal Pairs

Taiga Someya and 1 other author

How Much Syntactic Supervision is "Good Enough"?

Yohei Oseki and 1 other author

Context Limitations Make Neural Language Models More Human-Like

Tatsuki Kuribayashi and 3 other authors

Effective Batching for Recurrent Neural Network Grammars

Hiroshi Noji and 1 other author

Lower Perplexity is Not Always Human-Like

Tatsuki Kuribayashi and 5 other authors

Composition, Attention, or Both?

Ryo Yoshida and 1 other author

Stay up to date with the latest Underline news!

Select topic of interest (you can select more than one)

PRESENTATIONS

  • All Presentations
  • For Librarians
  • Resource Center
  • Free Trial
Underline Science, Inc.
1216 Broadway, 2nd Floor, New York, NY 10001, USA

© 2025 Underline - All rights reserved