Lecture image placeholder

Premium content

Access to this content requires a subscription. You must be a premium user to view this content.

Monthly subscription - $9.99Pay per view - $4.99Access through your institutionLogin with Underline account
Need help?
Contact us
Lecture placeholder background
VIDEO DOI: https://doi.org/10.48448/n4sa-wp38

poster

ACL 2024

August 12, 2024

Bangkok, Thailand

MORE: Multi-mOdal REtrieval Augmented Generative Commonsense Reasoning

keywords:

retrieval augmented generation

multi-modal

text generation

Since commonsense information has been recorded significantly less frequently than its existence, language models pre-trained by text generation have difficulty to learn sufficient commonsense knowledge. Several studies have leveraged text retrieval to augment the models' commonsense ability. Unlike text, images capture commonsense information inherently but little effort has been paid to effectively utilize them. In this work, we propose a novel $\textbf{M}$ulti-m$\textbf{O}$dal $\textbf{RE}$trieval (MORE) augmentation framework, to leverage both text and images to enhance the commonsense ability of language models. Extensive experiments on the Common-Gen task have demonstrated the efficacy of MORE based on the pre-trained models of both single and multiple modalities.

Downloads

SlidesTranscript English (automatic)

Next from ACL 2024

Learning Job Title Representation from Job Description Aggregation Network
poster

Learning Job Title Representation from Job Description Aggregation Network

ACL 2024

+2Ekapol Chuangsuwanich
Napat Laosaengpha and 4 other authors

12 August 2024

Stay up to date with the latest Underline news!

Select topic of interest (you can select more than one)

PRESENTATIONS

  • All Lectures
  • For Librarians
  • Resource Center
  • Free Trial
Underline Science, Inc.
1216 Broadway, 2nd Floor, New York, NY 10001, USA

© 2023 Underline - All rights reserved