Would you like to see your presentation here, made available to a global audience of researchers?
Add your own presentation or have us affordably record your next conference.
Multimodal retrieval models, such as Open-CLIP, rely on aligned textual and visual representations, yet their robustness to lexical variations remains underexplored, especially in low-resource languages. In this study, we introduce three methods for generating synonym substitutions in Ukrainian: a dictionary-based approach, a large language model (LLM)-generated approach, and a hybrid method combining both. We further demonstrate that fine-tuning Open-CLIP with synonym-augmented data improves retrieval robustness, leading to a 7% increase in HIT@5. Our findings provide insights into synonym substitution techniques for low-resource languages and offer a pathway to enhancing the robustness of multimodal models in diverse linguistic settings.