
Premium content
Access to this content requires a subscription. You must be a premium user to view this content.

workshop paper
Dolomites@#SMM4H 2024 Helping LLMs "Know The Drill" in Low-Resource Settings A Study on Social Media Posts
keywords:
mtl-da
entity recognition
large language model
classification
data augmentation
information extraction
The amount of data to fine-tune LLMs plays a crucial role in the performance of these models in downstream tasks. Consequently, it is not straightforward to deploy these models in lowresource settings. In this work, we investigate two new multi-task learning data augmentation approaches for fine-tuning LLMs when little data is available: "In-domain Augmentation" of the training data and extracting "Drills" as smaller tasks from the target dataset. We evaluate the proposed approaches in three natural language processing settings in the context of SMM4H 2024 competition tasks: multi-class classification, entity recognition, and information extraction. The results show that both techniques improve the performance of the models in all three settings, suggesting a positive impact from the knowledge learned in multi-task training to perform the target task.