profile picture

Anej Svete

Student @ ETH Zürich

language models

rnn

transformers

recurrent neural networks

finite-state automata

language model

prompt tuning

n-grams

linear temporal logic

failure transitions

semirings

pathsum

backward algorithm

chain of thought

formal language

8

presentations

SHORT BIO

Anej is a second-year PhD fellow at the ETH AI Center, where he is co-advised by prof. Ryan Cotterell and prof. Valentina Boeva.

His main research interests lie in the intersection of formal language theory and language models, where he is trying to understand the formal properties of architectures such as recurrent neural networks and transformers with weighted models of computation. He is also interested in representation learning and its interpretability.

Presentations

Can Transformer Language Models Learn $n$-gram Language Models?

Anej Svete and 4 other authors

On Efficiently Representing Regular Languages as RNNs

Anej Svete and 2 other authors

On the Representational Capacity of Neural Language Models with Chain-of-Thought Reasoning

Franz Nowak and 3 other authors

The Role of n-gram Smoothing in the Age of Neural Networks

Luca Malagutti and 5 other authors

Transformers Can Represent n-gram Language Models

Anej Svete and 1 other author

On the Relationship Between Non-deterministic FSLMs and RNN LMs

Anej Svete and 3 other authors

Recurrent Neural Language Models as Probabilistic Finite-state Automata

Anej Svete and 1 other author

On the Representational Capacity of Recurrent Neural Language Models

Franz Nowak and 3 other authors

Stay up to date with the latest Underline news!

Select topic of interest (you can select more than one)

PRESENTATIONS

  • All Presentations
  • For Librarians
  • Resource Center
  • Free Trial
Underline Science, Inc.
1216 Broadway, 2nd Floor, New York, NY 10001, USA

© 2025 Underline - All rights reserved