Source Themes

Training Plug-n-Play Knowledge Modules with Deep Context Distillation
Scaling Sparse Fine-Tuning to Large Language Models
Unifying Cross-Lingual Transfer across Scenarios of Resource Scarcity
Cross-Lingual Transfer with Target Language-Ready Task Adapters
Distilling Efficient Language-Specific Models for Cross-Lingual Transfer
Composable Sparse Fine-Tuning for Cross-Lingual Transfer
MAD-G: Multilingual Adapter Generation for Efficient Cross-Lingual Transfer
PolyLM: Learning about Polysemy through Language Modeling