
Anej Svete
Student @ ETH Zürich
language models
rnn
transformers
recurrent neural networks
finite-state automata
language model
prompt tuning
n-grams
linear temporal logic
failure transitions
semirings
pathsum
backward algorithm
chain of thought
formal language
8
presentations
SHORT BIO
Anej is a second-year PhD fellow at the ETH AI Center, where he is co-advised by prof. Ryan Cotterell and prof. Valentina Boeva.
His main research interests lie in the intersection of formal language theory and language models, where he is trying to understand the formal properties of architectures such as recurrent neural networks and transformers with weighted models of computation. He is also interested in representation learning and its interpretability.
Presentations

Can Transformer Language Models Learn $n$-gram Language Models?
Anej Svete and 4 other authors

On Efficiently Representing Regular Languages as RNNs
Anej Svete and 2 other authors

On the Representational Capacity of Neural Language Models with Chain-of-Thought Reasoning
Franz Nowak and 3 other authors

The Role of n-gram Smoothing in the Age of Neural Networks
Luca Malagutti and 5 other authors

Transformers Can Represent n-gram Language Models
Anej Svete and 1 other author

On the Relationship Between Non-deterministic FSLMs and RNN LMs
Anej Svete and 3 other authors

Recurrent Neural Language Models as Probabilistic Finite-state Automata
Anej Svete and 1 other author

On the Representational Capacity of Recurrent Neural Language Models
Franz Nowak and 3 other authors