This Challenge Interspeech 2026 is a shared task aimed at advancing automatic assessment of Modern Standard Arabic (MSA) pronunciation by leveraging computational methods to detect and diagnose pronunciation errors. The focus on MSA provides a standardized and well-defined context for evaluating Arabic pronunciation.
Participants will develop systems capable of detecting mispronunciations (e.g., substitution, deletion, or insertion of phonemes).
Design a model to detect and provide detailed feedback on mispronunciations in MSA speech. Users read vowelized sentences; the model predicts the spoken phoneme sequence and flags deviations. Evaluation is on the MSA-Test dataset with human‐annotated errors.
Figure: Overview of the Mispronunciation Detection Workflow
System shows a Reference Sentence plus its Reference Phoneme Sequence.
Example:
< y a t a H a d d a v u n n aa s u l l u g h a t a l E a r a b i y y a t a
User speaks; system captures and stores the audio waveform.
Model predicts the phoneme sequence—deviations from reference indicate mispronunciations.
Example of Mispronunciation:
< y a t a H a d d a v u n n aa s u l l u g h a t a l E a r a b i y y a t a< y a t a H a d d a s u n n aa s u l l u g h a t u l E a r a b i y y a t a
Here, v→s (substitution) represents a common pronunciation error.
The phoneme set employed in this work derives from a specialized phonetizer developed specifically for vowelized Modern Standard Arabic (MSA). It encompasses a comprehensive inventory of phonemes designed to capture essential phonetic and prosodic features, including stress, pausing, intonation, emphaticness, and gemination. Notably, gemination—the lengthening of consonant sounds—is explicitly represented by duplicating the consonant symbol (e.g., /b/ becomes /bb/). This approach ensures a detailed yet practical representation of speech sounds, which is critical for accurate mispronunciation detection.
To phonemize additional datasets or custom text using this standard, we provide the open-source tool at the MSA Phonetizer Repository. Important: This phonetizer requires the input Arabic text to be fully diacritized to ensure accurate phonetic transcription. For further details on the symbols used, please refer to the Phoneme Inventory.
To ensure robustness, our training strategy utilizes a mix of native speech (pseudo-labeled), synthetic mispronunciations, and real recorded errors.
Dataset: IqraEval/Iqra_train
Volume: ~79 hours (Train) + 3.4 hours (Dev)
This dataset consists of recordings from native MSA speakers. As these speakers are assumed to pronounce the text correctly, this subset is treated as "Golden" data using pseudo-labels.
Columns:
audio: The speech waveform.sentence: The original raw text.tashkeel_sentence: Fully diacritized text, generated using an internal SOTA diacritizer (assumed correct).phoneme_ref: The reference canonical phoneme sequence.phoneme_mis: The realized phoneme sequence.
phoneme_ref.
Dataset: IqraEval/Iqra_TTS
Volume: ~80 hours
To compensate for the lack of errors in the native set, we generated a synthetic dataset using various trained TTS systems. Mispronunciations were deliberately introduced into the input text before audio generation.
Columns:
audio: The synthesized waveform.sentence_ref: The original correct text.sentence_aug: The text containing augmented errors.phoneme_ref: The canonical phoneme sequence of the correct text.phoneme_mis: The phoneme sequence corresponding to the synthesized mispronunciation.tashkeel_sentence: The fully diacritized version of the reference text.
Dataset: IqraEval/Iqra_Extra_IS26
Volume: ~2 hours
Moving beyond synthetic data, this subset contains real recordings of human mispronunciations collected specifically for Interspeech 2026.
Columns:
audio: The speech waveform.sentence: The original text.phoneme_ref: The target canonical phoneme sequence.phoneme_mis: The actual realized phonemes containing human errors.
Dataset: IqraEval/QuranMB.v2
Currently, only the audio files are released for this evaluation set. It serves as a benchmark for detecting mispronunciations in a distinct domain.
The Leaderboard is ranked primarily by the Phoneme-level F1-score. While other metrics (FRR, FAR, DER) are computed for analysis, F1 determines the final standing.
We use a hierarchical evaluation strategy (detection + diagnostic) based on the MDD Overview framework.
From the inputs above, we compute the following counts:
| TA (True Accept) | Correct phonemes properly accepted. |
| TR (True Reject) | Mispronunciations correctly detected. |
| FR (False Reject) | Correct phonemes incorrectly flagged as errors. |
| FA (False Accept) | Mispronunciations missed (labeled as correct). |
Teams and individual participants must register to gain access to the test set. Please complete the registration form using the link below:
Registration opens on December 1, 2025.
Further details on the open-set leaderboard submission will be posted on the shared task website (December 15, 2025). Stay tuned!
For inquiries and support, reach out to the task coordinators.