Skip to content

sinaahmadi/PersoArabicLID

Repository files navigation

Language Identification for Perso-Arabic Scripts

Perso-Arabic scripts word cloud

The Perso-Arabic scripts are a family of scripts that are widely adopted and used by various linguistic communities around the globe. Identifying various languages using such scripts is crucial to language technologies and challenging in low-resource setups. As such, this paper sheds light on the challenges of detecting languages using Perso-Arabic scripts, especially in bilingual communities where “unconventional” writing is practiced. To address this, we use a set of supervised techniques to classify sentences into their languages. Building on these, we also propose a hierarchical model that targets clusters of languages that are more often confused by the classifiers.

This repository provides datasets and models for language identification with a focus on languages that use a Perso-Arabic script. The selected 19 languages with their ISO-639-3 codes are provided as follows:

  • Brahui / براہوئی (brh)
  • Torwali / توروالی (trw)
  • Balochi / بلۏچی (bal)
  • Kashmiri / كٲشُر (kas)
  • Gorani / گۆرانی (Hawrami, hac)
  • Northern Kurdish / کورمانجی (Kurmanji, kmr-arab)
  • Central Kurdish / سۆرانی (Sorani, ckb)
  • Southern Kurdish / کوردیی خوارین (sdh)
  • Arabic / اَلْعَرَبِيَّةُ (arb)
  • Persian / فارسی (fas)
  • Gilaki / گیلکی (glk)
  • Mazanderani / مازرونی (mzn)
  • Urdu / اردو (urd)
  • Sindhi / سنڌي (snd-arab)
  • Azeri Turkish / آذربایجان دیلی (azb-arab)
  • Uyghur / ئۇيغۇر تىلى (uig)
  • Pashto / پښتو (pas)
  • Saraiki / سرائیکی (skr)
  • Punjabi / پنجابی (pnb)

Corpora

The corpora folder contains corpora for the target languages.

Datasets

The datasets folder is structured as follows:

  1. Clean: A dataset of clean data (0% of noise) is provided in datasets/0. This includes all the selected languages along with Urdu, Persian, Arabic and Uyghur. The dataset contains 10,000 instances per language.
  2. Noisy: Based on the script mappings in scripts, a certain level of noise is injected in the clean text from the corpora of the source language. This is carried out in create_datasets.py. The following datasets containing 10,000 instances per language are provided based on the indicated level of noise from 20% to 100% as:
  3. Merged: Folder datasets/merged contains instances from the clean and noisy datasets (the two previous ones). To avoid imbalance of clean languages for which there is no noise, i.e. Urdu, Persian, Arabic and Uyghur, 10,000 more instances are added from their clean corpora. Therefore, there should be 20000 instances per language.

It should be noted that we upsample data for some of the languages for which there are < 10,000 sentences in our corpora. This is summarized as follows (# refers to the number of sentences):

Language # clean corpus Upsample (# clean train set) # Clean test set
Brahui 549 Yes (200) 500
Torwali 1371 Yes (2000) 500
Balochi 1649 Yes (2000) 500
Kashmiri 6340 Yes (8000) 2000
Gorani 8742 Yes (8000) 2000
others > 10,000 No 2000

Models

The models folder provides compressed models that are trained on the train sets. These models are trained on all the selected datasets.

You can load the trained models using fastText in Python or on command-line. For more information, please consult https://fasttext.cc/docs/en/python-module.html.

Here is an example in Python:

>>> import fasttext
>>> model = fasttext.load_model("LID_model_merged.ftz")
>>> model.predict("لەزۆربەی یارییەکان گوڵ تۆمار دەکات")
(('__label__ckb',), array([0.96654707]))

  • fa: Farsi
  • kas: Kashmiri
  • sd: Sindhi
  • azb: Azeri
  • ckb: Central Kurdish
  • bal: Balochi
  • mzn: Mazanderani
  • hac: Gorani
  • skr: Saraiki
  • ps: Pashto
  • ar: Arabic
  • trw: Torwali
  • glk: Gilaki
  • sdh: Sindhi
  • ud: Urdu
  • ug: Uyghur
  • pa: Punjabi
  • ku: Northern Kurdish
  • brh: Brahui

The predicted language code is preceded by __lable__. We recommend using the merged model (LID_model_merged.ftz).

Utilities

  • stats.sh counts number of instances per languages in the various datasets. It can be used with clean, merged or noisy as arguments. The latter should also take the noise level as an additional argument as in stats.sh noisy 20 or stats.sh noisy all.
  • merger.sh creates the merged datasets which performs equally well in noisy and clean setups.

Cite this project

Please consider citing this paper, if you use any part of the data or the models (bib file):

@inproceedings{ahmadi2023pali,
    title = "{PALI}: A Language Identification Benchmark for {Perso-Arabic} Scripts",
    author = "Ahmadi, Sina and Agarwal, Milind and Anastasopoulos, Antonios",
    booktitle = "Proceedings of the 10th Workshop on NLP for Similar Languages, Varieties and Dialects (VarDial)",
    month = may,
    year = "2023",
    address = "Dubrovnik, Croatia",
    publisher = "The 17th Conference of the European Chapter of the Association for Computational Linguistics"
}

License

MIT