Paper published in a book (Scientific congresses, symposiums and conference proceedings)
Enhancing Small Language Models for Cross-Lingual Generalized Zero-Shot Classification with Soft Prompt Tuning
PHILIPPY, Fred; Guo, Siwen; Lothritz, Cedric et al.
2025In Enhancing Small Language Models for Cross-Lingual Generalized Zero-Shot Classification with Soft Prompt Tuning
Peer reviewed
 

Files


Full Text
2025.lm4uc-1.8.pdf
Publisher postprint (885.77 kB) Creative Commons License - Attribution, ShareAlike
Download

All documents in ORBilu are protected by a user license.

Send to



Details



Keywords :
NLP, Low-Resource Languages, Parameter Efficiency
Abstract :
[en] In NLP, Zero-Shot Classification (ZSC) has become essential for enabling models to classify text into categories unseen during training, particularly in low-resource languages and domains where labeled data is scarce. While pretrained language models (PLMs) have shown promise in ZSC, they often rely on large training datasets or external knowledge, limiting their applicability in multilingual and low-resource scenarios.Recent approaches leveraging natural language prompts reduce the dependence on large training datasets but struggle to effectively incorporate available labeled data from related classification tasks, especially when these datasets originate from different languages or distributions. Moreover, existing prompt-based methods typically rely on manually crafted prompts in a specific language, limiting their adaptability and effectiveness in cross-lingual settings.To address these challenges, we introduce RoSPrompt, a lightweight and data-efficient approach for training soft prompts that enhance cross-lingual ZSC while ensuring robust generalization across data distribution shifts. RoSPrompt is designed for small multilingual PLMs, enabling them to leverage high-resource languages to improve performance in low-resource settings without requiring extensive fine-tuning or high computational costs. We evaluate our approach on multiple multilingual PLMs across datasets covering 106 languages, demonstrating strong cross-lingual transfer performance and robust generalization capabilities over unseen classes.
Disciplines :
Computer science
Author, co-author :
PHILIPPY, Fred  ;  University of Luxembourg
Guo, Siwen;  Zortify S.A.
Lothritz, Cedric;  LIST - Luxembourg Institute of Science and Technology
KLEIN, Jacques  ;  University of Luxembourg > Interdisciplinary Centre for Security, Reliability and Trust (SNT) > TruX
BISSYANDE, Tegawendé François d Assise  ;  University of Luxembourg > Interdisciplinary Centre for Security, Reliability and Trust (SNT) > TruX
External co-authors :
no
Language :
English
Title :
Enhancing Small Language Models for Cross-Lingual Generalized Zero-Shot Classification with Soft Prompt Tuning
Publication date :
May 2025
Event name :
1st Workshop on Language Models for Underserved Communities (LM4UC 2025)
Event place :
Albuquerque, United States - New Mexico
Event date :
May 2025
Audience :
International
Main work title :
Enhancing Small Language Models for Cross-Lingual Generalized Zero-Shot Classification with Soft Prompt Tuning
Publisher :
Association for Computational Linguistics
Peer reviewed :
Peer reviewed
Focus Area :
Computational Sciences
Available on ORBilu :
since 06 June 2025

Statistics


Number of views
65 (9 by Unilu)
Number of downloads
23 (1 by Unilu)

OpenCitations
 
0
OpenAlex citations
 
1

Bibliography


Similar publications



Contact ORBilu