{"id":72894,"date":"2026-02-28T15:29:16","date_gmt":"2026-02-28T20:29:16","guid":{"rendered":"https:\/\/languagelog.ldc.upenn.edu\/nll\/?p=72894"},"modified":"2026-02-28T19:16:17","modified_gmt":"2026-03-01T00:16:17","slug":"unifying-arabic-topolects-through-ai","status":"publish","type":"post","link":"https:\/\/languagelog.ldc.upenn.edu\/nll\/?p=72894","title":{"rendered":"Unifying Arabic topolects through AI"},"content":{"rendered":"<p style=\"padding-left: 40px;\"><a href=\"https:\/\/www.scmp.com\/news\/china\/article\/3344955\/meet-habibi-chinese-ai-uniting-20-arabic-dialects-middle-east-first\">Meet Habibi \u2013 the Chinese AI unit<\/a><a href=\"https:\/\/www.scmp.com\/news\/china\/article\/3344955\/meet-habibi-chinese-ai-uniting-20-arabic-dialects-middle-east-first\">ing 20 Arabic dialects in a Middle East first<\/a><br \/>Lead author says there are many differences between Arabic dialects and Modern Standard Arabic, which is used in official circumstances<br \/>Zhao Ziwen, SCMP, 28 Feb 2026<\/p>\r\n<p>The paper that presents this new model is called\u00a0\u201cHabibi: Laying the Open-Source Foundation of Unified-Dialectal Arabic Speech Synthesis\u201d. It was published last month on arXiv, an open-access repository that is not peer-reviewed.\u00a0 I will be interested to hear what Language Log readers think of its prospects.<\/p>\r\n<p><!--more--><\/p>\r\n<div class=\"e8zc9q40 css-1xdhyk6 ec74h0k0\" style=\"padding-left: 40px;\" data-qa=\"Component-Component\">Chinese researchers have released the world\u2019s first open-source text-to-speech (TTS) model that unifies more than 20 Arabic dialects in an <a class=\"e1yy41x40 ef9u0v01 css-1ankfgb ecgc78b0\" title=\"\" href=\"https:\/\/www.scmp.com\/topics\/artificial-intelligence?module=inline&amp;pgtype=article\" target=\"_self\" data-qa=\"BaseLink-renderAnchor-StyledAnchor\"><span class=\"css-0 ef9u0v00\" data-qa=\"Component-Text\">AI framework<\/span><\/a>, a move poised to expand <a class=\"e1yy41x40 ef9u0v01 css-1ankfgb ecgc78b0\" title=\"\" href=\"https:\/\/www.scmp.com\/topics\/china-middle-east-relations?module=inline&amp;pgtype=article\" target=\"_self\" data-qa=\"BaseLink-renderAnchor-StyledAnchor\"><span class=\"css-0 ef9u0v00\" data-qa=\"Component-Text\">China\u2019s technological influence in the Middle East<\/span><\/a>, according to analysts.<\/div>\r\n<p class=\"e8zc9q40 css-1c6uqr6 ec74h0k1\" style=\"padding-left: 40px;\" data-qa=\"Component-Component\">Led by Shanghai Jiao Tong University\u2019s X-LANCE Lab \u2013 one of China\u2019s top audiovisual and language processing research entities \u2013 the model is named Habibi, meaning \u201cmy dear\u201d in Arabic.<\/p>\r\n<p class=\"e8zc9q40 css-1c6uqr6 ec74h0k1\" style=\"padding-left: 40px;\" data-qa=\"Component-Component\">In presenting their findings, the research team spearheaded by Chen Yushen described the project in a paper as \u201cthe first open-source framework for unified-dialectal Arabic speech synthesis\u201d.<\/p>\r\n<p>They introduce a concept that is new to me:\u00a0 \u00a0\"zero-shot\".<\/p>\r\n<p style=\"padding-left: 40px;\">Habibi has the \u201czero-shot\u201d ability, meaning the model can easily clone a voice by using just a short reference audio clip, without prior explicit or extensive training. This allows applications in highly efficient and on-the-fly scenarios.<\/p>\r\n<p>According to <a href=\"https:\/\/en.wikipedia.org\/wiki\/Zero-shot_learning\">Wikipedia<\/a>,<\/p>\r\n<p style=\"padding-left: 40px;\"><b>Zero-shot learning<\/b> (<b>ZSL<\/b>) is a problem setup in <a title=\"Deep learning\" href=\"https:\/\/en.wikipedia.org\/wiki\/Deep_learning\">deep learning<\/a> where, at test time, a learner observes samples from classes which were <i>not<\/i> observed during <a title=\"Machine learning\" href=\"https:\/\/en.wikipedia.org\/wiki\/Machine_learning#Training_models\">training<\/a>, and needs to predict the class that they belong to. The name is a play on words based on the earlier concept of <a class=\"mw-redirect\" title=\"One-shot learning in computer vision\" href=\"https:\/\/en.wikipedia.org\/wiki\/One-shot_learning_in_computer_vision\">one-shot learning<\/a>, in which classification can be learned from only one, or a few, examples.<\/p>\r\n<p style=\"padding-left: 40px;\">Zero-shot methods generally work by associating observed and non-observed classes through some form of auxiliary information, which encodes observable distinguishing properties of objects.\u00a0\u00a0For example, given a set of images of animals to be classified, along with auxiliary textual descriptions of what animals look like, an artificial intelligence model which has been trained to recognize horses, but has never been given a zebra, can still recognize a zebra when it also knows that zebras look like striped horses. This problem is widely studied in <a title=\"Computer vision\" href=\"https:\/\/en.wikipedia.org\/wiki\/Computer_vision\">computer vision<\/a>, <a title=\"Natural language processing\" href=\"https:\/\/en.wikipedia.org\/wiki\/Natural_language_processing\">natural language processing<\/a>, and <a title=\"Machine perception\" href=\"https:\/\/en.wikipedia.org\/wiki\/Machine_perception\">machine perception<\/a>.<\/p>\r\n<div id=\"bodyContent\" class=\"vector-body ve-init-mw-desktopArticleTarget-targetContainer\" style=\"padding-left: 40px;\" aria-labelledby=\"firstHeading\" data-mw-ve-target-container=\"\">\r\n<div id=\"mw-content-text\" class=\"mw-body-content\" style=\"padding-left: 40px;\">\r\n<div class=\"mw-content-ltr mw-parser-output\" dir=\"ltr\" lang=\"en\" style=\"padding-left: 40px;\">\r\n<figure class=\"mw-default-size\"><a class=\"mw-file-description\" href=\"https:\/\/en.wikipedia.org\/wiki\/File:Nature_of_Ngorongoro_Conservation_Area_(23).jpg\"><img loading=\"lazy\" decoding=\"async\" class=\"mw-file-element\" src=\"https:\/\/upload.wikimedia.org\/wikipedia\/commons\/thumb\/a\/ac\/Nature_of_Ngorongoro_Conservation_Area_%2823%29.jpg\/250px-Nature_of_Ngorongoro_Conservation_Area_%2823%29.jpg\" width=\"250\" height=\"167\" data-file-width=\"6720\" data-file-height=\"4480\" \/><\/a>\r\n<figcaption>A zebra can be identified as looking like a striped horse, even if you've never seen a zebra before<br \/><br \/><\/figcaption>\r\n<\/figure>\r\n<\/div>\r\n<\/div>\r\n<\/div>\r\n<p><b>Selected readings<\/b><\/p>\r\n<ul>\r\n<li>\"<a title=\"Permanent link to LLMs and tree-structuring\" href=\"https:\/\/languagelog.ldc.upenn.edu\/nll\/?p=71128\" rel=\"bookmark\">LLMs and tree-structuring<\/a>\" (9\/18\/25)<\/li>\r\n<li>\"<a title=\"Permanent link to Radial dendrograms\" href=\"https:\/\/languagelog.ldc.upenn.edu\/nll\/?p=59863\" rel=\"bookmark\">Radial dendrograms<\/a>\u00a0(7\/26\/23)<\/li>\r\n<li>\"<a title=\"Permanent link to Language trees and script trees\" href=\"https:\/\/languagelog.ldc.upenn.edu\/nll\/?p=53144\" rel=\"bookmark\">Language trees and script trees<\/a>\" (12\/27\/21)<\/li>\r\n<li>\"<a title=\"Permanent link to AMI not AGI?\" href=\"https:\/\/languagelog.ldc.upenn.edu\/nll\/?p=70255\" rel=\"bookmark\">AMI not AGI?<\/a>\" (8\/2\/25)<\/li>\r\n<\/ul>\r\n<p><b>Addendum<\/b><\/p>\r\n<p>In case you're interested, \"Habibi\" itself is an Arabic word worth learning in one of its 20 plus topolects:\u00a0 Syrian, Egyptian, Jordanian, Levantine&#8230;.\u00a0 \u00a0Because of its wide range of meanings, nuances, and usages, be careful of how, when, and to whom you use it.<\/p>\r\n<p>Listen <a href=\"https:\/\/playaling.com\/habibi-in-arabic\/\">here<\/a>.<\/p>\r\n<p>[Thanks to Mark Metcalf]<\/p>\r\n","protected":false},"excerpt":{"rendered":"<p>Meet Habibi \u2013 the Chinese AI uniting 20 Arabic dialects in a Middle East firstLead author says there are many differences between Arabic dialects and Modern Standard Arabic, which is used in official circumstancesZhao Ziwen, SCMP, 28 Feb 2026 The paper that presents this new model is called\u00a0\u201cHabibi: Laying the Open-Source Foundation of Unified-Dialectal Arabic [&hellip;]<\/p>\n","protected":false},"author":13,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_exactmetrics_skip_tracking":false,"_exactmetrics_sitenote_active":false,"_exactmetrics_sitenote_note":"","_exactmetrics_sitenote_category":0,"jetpack_post_was_ever_published":false,"_jetpack_newsletter_access":"","_jetpack_dont_email_post_to_subs":false,"_jetpack_newsletter_tier_id":0,"_jetpack_memberships_contains_paywalled_content":false,"_jetpack_memberships_contains_paid_content":false,"footnotes":""},"categories":[322,60,124,224],"tags":[],"class_list":["post-72894","post","type-post","status-publish","format-standard","hentry","category-artificial-intelligence","category-computational-linguistics","category-dialects","category-topolects"],"jetpack_featured_media_url":"","jetpack_sharing_enabled":true,"_links":{"self":[{"href":"https:\/\/languagelog.ldc.upenn.edu\/nll\/index.php?rest_route=\/wp\/v2\/posts\/72894","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/languagelog.ldc.upenn.edu\/nll\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/languagelog.ldc.upenn.edu\/nll\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/languagelog.ldc.upenn.edu\/nll\/index.php?rest_route=\/wp\/v2\/users\/13"}],"replies":[{"embeddable":true,"href":"https:\/\/languagelog.ldc.upenn.edu\/nll\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=72894"}],"version-history":[{"count":5,"href":"https:\/\/languagelog.ldc.upenn.edu\/nll\/index.php?rest_route=\/wp\/v2\/posts\/72894\/revisions"}],"predecessor-version":[{"id":72933,"href":"https:\/\/languagelog.ldc.upenn.edu\/nll\/index.php?rest_route=\/wp\/v2\/posts\/72894\/revisions\/72933"}],"wp:attachment":[{"href":"https:\/\/languagelog.ldc.upenn.edu\/nll\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=72894"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/languagelog.ldc.upenn.edu\/nll\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=72894"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/languagelog.ldc.upenn.edu\/nll\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=72894"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}