
Premium content
Access to this content requires a subscription. You must be a premium user to view this content.

workshop paper
Adapting transformer models to morphological tagging of two highly inflectional languages: a case study on Ancient Greek and Latin
keywords:
inflectional languages
morphological tagging
ancient greek
latin
transformers
Natural language processing for Greek and Latin, inflectional languages with small corpora, requires special techniques. For morphological tagging, transformer models show promising potential, but the best approach to use these models is unclear. For both languages, this paper examines the impact of using morphological lexica, training different model types (a single model with a combined feature tag, multiple models for separate features, and a multi-task model for all features), and adding linguistic constraints. We find that, although simply fine-tuning transformers to predict a monolithic tag may already yield decent results, each of these adaptations can further improve tagging accuracy.