Dialectal Language Models: Challenges and Opportunities
<p>However, there’s been a surge of new Bilingual Language Learning Models, such as <em>CroissantLLM</em> for French, <em>Jais</em> for Arabic, and the <em>Japanese Stable LM</em> for Japanese. These specialized models tend to excel in handling their intended languages. Consequently, the emergence of dialect-focused LLMs raises intriguing questions — will future models master regional dialects?</p>
<p>In 2020, <a href="https://arxiv.org/pdf/2005.00318.pdf" rel="noopener ugc nofollow" target="_blank">Benjamin et al.</a> demonstrated that <em>mBERT</em> (multilingual <em>BERT</em>, pretrained on 104 Wikipedia languages) successfully transferred skills to <em>Narabizi</em>, an online Arabic dialect spoken in certain North African nations, even though it wasn’t included in the pretraining corpus. <strong>But what if we aim to develop a language model actually pre-trained on dialectal text data?</strong></p>
<p><a href="https://medium.com/@simosbaihi/dialectal-language-models-challenges-and-opportunities-ecd566491c4a"><strong>Learn More</strong></a></p>