Lecture image placeholder

Premium content

Access to this content requires a subscription. You must be a premium user to view this content.

Monthly subscription - $9.99Pay per view - $4.99Access through your institutionLogin with Underline account
Need help?
Contact us
Lecture placeholder background

EMNLP 2021

November 08, 2021

Live on Underline

Would you like to see your presentation here, made available to a global audience of researchers?
Add your own presentation or have us affordably record your next conference.

Representation learning for text via pretraining a language model on a large corpus has become a standard starting point for building NLP systems. This approach stands in contrast to autoencoders, also trained on raw text, but with the objective of learning to encode each input as a vector that allows full reconstruction. Autoencoders are attractive because of their latent space structure and generative properties. We therefore explore the construction of a sentence-level autoencoder from a pretrained, frozen transformer language model. We adapt the masked language modeling objective as a generative, denoising one, while only training a sentence bottleneck and a single-layer modified transformer decoder. We demonstrate that the sentence representations discovered by our model achieve better quality than previous methods that extract representations from pretrained transformers on text similarity tasks, style transfer (an example of controlled generation), and single-sentence classification tasks in the GLUE benchmark, while using fewer parameters than large pretrained models.

Downloads

SlidesPaper
access premium content

Next from EMNLP 2021

Effects of Parameter Norm Growth During Transformer Training: Inductive Bias from Gradient Descent
technical paper

Effects of Parameter Norm Growth During Transformer Training: Inductive Bias from Gradient Descent

EMNLP 2021

+2Yoav GoldbergWilliam Merrill
William Merrill and 4 other authors

08 November 2021

Similar lecture

One-Shot Lexicon Learning for Low-Resource Machine Translation
workshop paper

One-Shot Lexicon Learning for Low-Resource Machine Translation

EMNLP 2021

Jacob AndreasAnjali Kantharuban
Anjali Kantharuban and 1 other author

08 November 2021

Stay up to date with the latest Underline news!

Select topic of interest (you can select more than one)

PRESENTATIONS

  • All Presentations
  • For Librarians
  • Resource Center
  • Free Trial
Underline Science, Inc.
1216 Broadway, 2nd Floor, New York, NY 10001, USA

© 2026 Underline - All rights reserved