Skip to main content
U.S. flag

An official website of the United States government

Official websites use .gov
A .gov website belongs to an official government organization in the United States.

Secure .gov websites use HTTPS
A lock ( ) or https:// means you’ve safely connected to the .gov website. Share sensitive information only on official, secure websites.

OpenASR21 Challenge Results

The OpenASR (Open Automatic Speech Recognition) 2021 Challenge was the third open challenge and the second OpenASR challenge associated with the IARPA MATERIAL program, after the OpenCLIR (Open Cross-Language Information Retrieval) 2019 Challenge and the OpenASR 2020 Challenge Capabilities tested in these open challenges are expected to ultimately support the MATERIAL task of effective triage and analysis of large volumes of text and audio content in a variety of less-studied languages. OpenASR21 was implemented as a track of NIST’s OpenSAT (Speech Analytic Technologies) evaluation series.

The goal of the OpenASR21 challenge was to assess the state of the art of automatic speech recognition (ASR) technologies for low-resource languages. ASR was performed on speech datasets, and written text output had to be produced.

Please refer to the OpenASR21 Challenge Evaluation Plan for a full description of the challenge and its rules and procedures.

Languages and Casing

OpenASR21 was offered for 15 low-resource languages, of which participating teams could attempt as many as they wished. Ten of these languages were repeated from OpenASR20 with unchanged datasets; five languages were new. All languages were offered with a case-insensitive scoring dataset. Of the five new languages, three were offered with an additional case-sensitive scoring dataset. The case-sensitive scoring was to serve as a proxy for assessment of ASR performance on proper nouns. The new languages are marked as such below:

  • Amharic (AMH)
  • Cantonese (CAN)
  • New: Georgian (GEO)
  • New: Farsi (FAR)
  • Guarani (GUA)
  • Javanese (JAV)
  • New: Kazakh (KAZ) (including additional case-sensitive scoring dataset)
  • Kurmanji Kurdish (KUR)
  • Mongolian (MON)
  • Pashto (PAS)
  • Somali (SOM)
  • New: Swahili (SWA) (including additional case-sensitive scoring dataset)
  • New: Tagalog (TAG) (including additional case-sensitive scoring dataset)
  • Tamil (TAM)
  • Vietnamese (VIE)

Data

The data for the challenge consisted of speech data stemming from the IARPA Babel program, with the exception of Somali. which stemmed from the IARPA MATERIAL program. More details regarding technical data details can be found in section 3 of the IARPA Babel Data Specifications for Performers. For each language (and case scoring condition, where applicable), separate training, development, and evaluation datasets were provided.

For CIS, the datasets for most of the languages stemmed from the IARPA Babel program. The Somali and Farsi datasets stemmed from the IARPA MATERIAL program. For the case-sensitive scoring datasets, all data stemmed from the IARPA MATERIAL program.

The data consisted of conversational speech for the case-insensitive scoring datasets. For the case-sensitive scoring datasets, the data consisted of a mix of three genres: conversational speech, news broadcast, and topical broadcast.

Training Conditions

The challenge offered three training conditions. The Constrained-plus training was new for OpenASR21.

  • Constrained training (mandatory):
    • Speech data: Limited to a 10-hour subset designated for Constrained training in provided training dataset for the language in question
    • Non-speech data: Any publicly available data
  • New: Constrained-plus training (optional):
    • Same training data restrictions as Constrained Training, but additionally allowed publicly available and previously existing speech pretrained models as follows:
      • Pretrained models created from unlabeled speech data in any language
      • Pretrained models created from labeled speech data in any language except the language being processed
  • Unconstrained training (optional):
    • Any publicly available data

Metrics

  • Primary metric: Word Error Rate (WER)
  • Additional metrics:
    • Character Error Rate (CER)
    • Time and memory resources used (self-reported)

Schedule

The most important milestones of the schedule of the challenge were as follows:

  • Registration: August - October, 2021
  • Evaluation period: November 3-10, 2021

Participation

Table 1 lists the teams that submitted valid output, and for which languages they submitted valid output.

26 teams from 13 countries originally registered to participate. Fifteen teams submitted valid output for at least one language under the Constrained training condition.

The table marks the following issues in the pertinent table cells:

  • The team only made late submissions (up to seven days after the official evaluation period ended) for the language in question.
  • The team did not make a required Constrained training condition submission for the language in question.
  • The team did not submit the required system description.
Organization Team AMH CAN FAR GEO GUA JAV KAZ KUR MON PAS SOM SWA TAG TAM VIE
Shanghai Jian Qiao University, China Baymax (no system description) Yes (no Constrained submission)                       Yes    
Catskills Research Company, USA Catskills     Yes               Yes        
Deep Learning and Media System Laboratory, National Central University, Taiwan, China CHNGA-DLMSL (no system description)   Yes Yes (late)                        
Centre de Recherche Informatique de Montréal, Canada CRIM Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes
Northeastern University, China FiberHome-AILab (no system description)   Yes                          
Institute for Infocomm Research, Singapore I2R           Yes             Yes Yes (late) Yes
University of Exeter, UK Jarvs (no system description) Yes (no Constrained submission)                            
Tencent, China MMT   Yes (no Constrained submission)         Yes (no Constrained submission)   Yes            
Samsung R&D Institute Bangalore, India SRIB_ASR                           Yes  
Tallinn University of Technology, Estonia TalTech Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes
Tsinghua University, China THUEE Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes
Tencent & Tsinghua University, China TNT   Yes         Yes (no Constrained submission)   Yes            
Transsion AI Technology Dept., China TranSpeech (no system description) Yes (no Constrained submission)                     Yes (late, no Constrained submission) Yes    
National Engineering Laboratory for Speech and Language Information Processing, University of Science and Technology of China ustc_nelslip Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes
SpeakIn Technologies Co., Ltd, China YCG-AI (no system description)   Yes                          
University of Science and Technology of China zxy Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes

Table 1: OpenASR21 Participating teams and languages.

Results

Table 2 lists the best WER result achieved by each team, ordered by language, scoring, training condition, and WER score. The CER score for the same submission is also provided.

The table marks the following issues in the Notes column:

  • The submission was late (up to seven days after the official evaluation period ended). For some cases, this results in two best WER scores shown for a given team, language, and conditions - their best score, which was for a late submission, as well as their best on-time score. Late submissions are listed at the bottom of the table.
  • The team did not make a required Constrained training condition submission for the language in question.
  • The team did not submit the required system description.

Self-reported time and memory resources are not included in this overview of results.

On-time submissions
Language Scoring Training Team WER CER Notes
Amharic Case-insensitive Constrained ustc_nelslip 0.3993 0.3019  
Amharic Case-insensitive Constrained THUEE 0.4219 0.3331  
Amharic Case-insensitive Constrained zxy 0.4257 0.3114  
Amharic Case-insensitive Constrained CRIM 0.4310 0.3288  
Amharic Case-insensitive Constrained TalTech 0.4318 0.3277  
Language Scoring Training Team WER CER Notes
Amharic Case-insensitive Constrained-plus THUEE 0.3729 0.2753  
Amharic Case-insensitive Constrained-plus zxy 0.4045 0.2986  
Amharic Case-insensitive Constrained-plus TalTech 0.4152 0.3101  
Language Scoring Training Team WER CER Notes
Amharic Case-insensitive Unconstrained TalTech 0.3522 0.2609  
Amharic Case-insensitive Unconstrained TranSpeech 0.4424 0.3167 No Constrained submission, no system description
Amharic Case-insensitive Unconstrained Jarvs 0.4446 0.3191 No Constrained submission, no system description
Amharic Case-insensitive Unconstrained Baymax 0.5100 0.3044 No Constrained submission, no system description
Language Scoring Training Team WER CER Notes
Cantonese Case-insensitive Constrained ustc_nelslip 0.3762 0.3274  
Cantonese Case-insensitive Constrained TNT 0.4024 0.3511  
Cantonese Case-insensitive Constrained THUEE 0.4038 0.3559  
Cantonese Case-insensitive Constrained TalTech 0.4267 0.3774  
Cantonese Case-insensitive Constrained CRIM 0.4277 0.3717  
Cantonese Case-insensitive Constrained zxy 0.4830 0.4245  
Cantonese Case-insensitive Constrained CHNGA-DLMSL 0.7113 0.6571 No system description
Cantonese Case-insensitive Constrained YCG-AI 0.7547 0.6669 No system description
Cantonese Case-insensitive Constrained FiberHome-AILab 1.2414 1.2284 No system description
Language Scoring Training Team WER CER Notes
Cantonese Case-insensitive Constrained-plus TNT 0.3371 0.2847  
Cantonese Case-insensitive Constrained-plus MMT 0.3631 0.3011 No Constrained submission
Cantonese Case-insensitive Constrained-plus zxy 0.3880 0.3337  
Cantonese Case-insensitive Constrained-plus TalTech 0.4101 0.3608  
Cantonese Case-insensitive Constrained-plus CHNGA-DLMSL 0.6834 0.6003 No system description
Cantonese Case-insensitive Constrained-plus YCG-AI 0.7091 0.5903 No system description
Language Scoring Training Team WER CER Notes
Cantonese Case-insensitive Unconstrained ustc_nelslip 0.2665 0.1982  
Cantonese Case-insensitive Unconstrained TNT 0.2776 0.2239  
Cantonese Case-insensitive Unconstrained MMT 0.2864 0.2332 No Constrained submission
Cantonese Case-insensitive Unconstrained zxy 0.3019 0.2238  
Cantonese Case-insensitive Unconstrained YCG-AI 0.7710 0.6246 No system description
Language Scoring Training Team WER CER Notes
Farsi Case-insensitive Constrained ustc_nelslip 0.6796 0.5472  
Farsi Case-insensitive Constrained zxy 0.6986 0.5366  
Farsi Case-insensitive Constrained CRIM 0.7932 0.6660  
Farsi Case-insensitive Constrained THUEE 0.8144 0.7533  
Farsi Case-insensitive Constrained TalTech 0.8159 0.7533  
Farsi Case-insensitive Constrained Catskills 0.9865 0.8898  
Language Scoring Training Team WER CER Notes
Farsi Case-insensitive Constrained-plus THUEE 0.6262 0.4978  
Farsi Case-insensitive Constrained-plus zxy 0.6980 0.5800  
Farsi Case-insensitive Constrained-plus TalTech 0.7906 0.7261  
Language Scoring Training Team WER CER Notes
Farsi Case-insensitive Unconstrained ustc_nelslip 0.5200 0.3691  
Farsi Case-insensitive Unconstrained zxy 0.5200 0.3691  
Language Scoring Training Team WER CER Notes
Georgian Case-insensitive Constrained ustc_nelslip 0.3918 0.2526  
Georgian Case-insensitive Constrained TalTech 0.4155 0.2632  
Georgian Case-insensitive Constrained zxy 0.4231 0.2665  
Georgian Case-insensitive Constrained CRIM 0.4255 0.2691  
Georgian Case-insensitive Constrained THUEE 0.4657 0.3296  
Language Scoring Training Team WER CER Notes
Georgian Case-insensitive Constrained-plus TalTech 0.3894 0.2415  
Georgian Case-insensitive Constrained-plus zxy 0.3978 0.2504  
Georgian Case-insensitive Constrained-plus THUEE 0.4039 0.2552  
Language Scoring Training Team WER CER Notes
Guarani Case-insensitive Constrained ustc_nelslip 0.4262 0.3939  
Guarani Case-insensitive Constrained THUEE 0.4423 0.4099  
Guarani Case-insensitive Constrained TalTech 0.4524 0.4168  
Guarani Case-insensitive Constrained zxy 0.4548 0.4208  
Guarani Case-insensitive Constrained CRIM 0.4597 0.4297  
Language Scoring Training Team WER CER Notes
Guarani Case-insensitive Constrained-plus THUEE 0.4058 0.3706  
Guarani Case-insensitive Constrained-plus zxy 0.4312 0.3954  
Guarani Case-insensitive Constrained-plus TalTech 0.4375 0.4034  
Language Scoring Training Team WER CER Notes
Javanese Case-insensitive Constrained ustc_nelslip 0.4808 0.4813  
Javanese Case-insensitive Constrained zxy 0.5110 0.5115  
Javanese Case-insensitive Constrained CRIM 0.5197 0.5201  
Javanese Case-insensitive Constrained THUEE 0.5210 0.5216  
Javanese Case-insensitive Constrained TalTech 0.5272 0.5276  
Javanese Case-insensitive Constrained I2R 0.8957 0.8961  
Language Scoring Training Team WER CER Notes
Javanese Case-insensitive Constrained-plus THUEE 0.4797 0.4801  
Javanese Case-insensitive Constrained-plus zxy 0.4838 0.4845  
Javanese Case-insensitive Constrained-plus TalTech 0.5058 0.5063  
Language Scoring Training Team WER CER Notes
Javanese Case-insensitive Unconstrained ustc_nelslip 0.3951 0.3920  
Javanese Case-insensitive Unconstrained zxy 0.4442 0.4450  
Language Scoring Training Team WER CER Notes
Kazakh Case-insensitive Constrained ustc_nelslip 0.4997 0.3794  
Kazakh Case-insensitive Constrained TalTech 0.5300 0.3817  
Kazakh Case-insensitive Constrained zxy 0.5334 0.3696  
Kazakh Case-insensitive Constrained THUEE 0.5454 0.4373  
Kazakh Case-insensitive Constrained CRIM 0.5685 0.4378  
Language Scoring Training Team WER CER Notes
Kazakh Case-insensitive Constrained-plus TNT 0.4279 0.2903 No Constrained submission
Kazakh Case-insensitive Constrained-plus THUEE 0.4287 0.2889  
Kazakh Case-insensitive Constrained-plus zxy 0.4991 0.3606  
Kazakh Case-insensitive Constrained-plus TalTech 0.5058 0.3646  
Language Scoring Training Team WER CER Notes
Kazakh Case-insensitive Unconstrained ustc_nelslip 0.3745 0.2499  
Kazakh Case-insensitive Unconstrained zxy 0.3812 0.2545  
Kazakh Case-insensitive Unconstrained TNT 0.3945 0.2928 No Constrained submission
Kazakh Case-insensitive Unconstrained MMT 0.4036 0.2928 No Constrained submission
Language Scoring Training Team WER CER Notes
Kazakh Case-sensitive Constrained THUEE 0.4983 0.2900  
Kazakh Case-sensitive Constrained zxy 0.5468 0.2931  
Kazakh Case-sensitive Constrained CRIM 0.5855 0.3206  
Language Scoring Training Team WER CER Notes
Kazakh Case-sensitive Constrained-plus THUEE 0.4989 0.3282  
Kazakh Case-sensitive Constrained-plus zxy 0.5285 0.3026  
Language Scoring Training Team WER CER Notes
Kurmanji-Kurdish Case-insensitive Constrained ustc_nelslip 0.6168 0.5776  
Kurmanji-Kurdish Case-insensitive Constrained TalTech 0.6437 0.6079  
Kurmanji-Kurdish Case-insensitive Constrained zxy 0.6537 0.6076  
Kurmanji-Kurdish Case-insensitive Constrained CRIM 0.6567 0.6173  
Kurmanji-Kurdish Case-insensitive Constrained THUEE 0.6585 0.6272  
Language Scoring Training Team WER CER Notes
Kurmanji-Kurdish Case-insensitive Constrained-plus THUEE 0.6154 0.5673  
Kurmanji-Kurdish Case-insensitive Constrained-plus zxy 0.6185 0.5809  
Kurmanji-Kurdish Case-insensitive Constrained-plus TalTech 0.6213 0.5846  
Language Scoring Training Team WER CER Notes
Mongolian Case-insensitive Constrained ustc_nelslip 0.4098 0.3025  
Mongolian Case-insensitive Constrained MMT 0.4165 0.3048  
Mongolian Case-insensitive Constrained TNT 0.4171 0.3006  
Mongolian Case-insensitive Constrained zxy 0.4489 0.3169  
Mongolian Case-insensitive Constrained TalTech 0.4531 0.3343  
Mongolian Case-insensitive Constrained THUEE 0.4531 0.3530  
Mongolian Case-insensitive Constrained CRIM 0.4603 0.3357  
Language Scoring Training Team WER CER Notes
Mongolian Case-insensitive Constrained-plus TNT 0.3781 0.2824  
Mongolian Case-insensitive Constrained-plus MMT 0.3787 0.2825  
Mongolian Case-insensitive Constrained-plus THUEE 0.4113 0.3002  
Mongolian Case-insensitive Constrained-plus zxy 0.4190 0.3047  
Mongolian Case-insensitive Constrained-plus TalTech 0.4366 0.3211  
Language Scoring Training Team WER CER Notes
Mongolian Case-insensitive Unconstrained ustc_nelslip 0.3152 0.2509  
Mongolian Case-insensitive Unconstrained zxy 0.3188 0.2514  
Mongolian Case-insensitive Unconstrained TNT 0.3430 0.2525  
Mongolian Case-insensitive Unconstrained MMT 0.3728 0.2781  
Language Scoring Training Team WER CER Notes
Pashto Case-insensitive Constrained ustc_nelslip 0.4322 0.3045  
Pashto Case-insensitive Constrained TalTech 0.4530 0.3236  
Pashto Case-insensitive Constrained THUEE 0.4639 0.3302  
Pashto Case-insensitive Constrained CRIM 0.4723 0.3256  
Pashto Case-insensitive Constrained zxy 0.4734 0.3122  
Language Scoring Training Team WER CER Notes
Pashto Case-insensitive Constrained-plus THUEE 0.4138 0.2812  
Pashto Case-insensitive Constrained-plus TalTech 0.4249 0.2950  
Pashto Case-insensitive Constrained-plus zxy 0.4401 0.2993  
Language Scoring Training Team WER CER Notes
Pashto Case-insensitive Unconstrained ustc_nelslip 0.3388 0.2288  
Pashto Case-insensitive Unconstrained zxy 0.3403 0.2301  
Language Scoring Training Team WER CER Notes
Somali Case-insensitive Constrained ustc_nelslip 0.5560 0.5570  
Somali Case-insensitive Constrained THUEE 0.5802 0.5810  
Somali Case-insensitive Constrained zxy 0.5855 0.5866  
Somali Case-insensitive Constrained TalTech 0.5871 0.5879  
Somali Case-insensitive Constrained CRIM 0.5923 0.5933  
Somali Case-insensitive Constrained Catskills 0.9931 0.9931  
Language Scoring Training Team WER CER Notes
Somali Case-insensitive Constrained-plus THUEE 0.5334 0.5345  
Somali Case-insensitive Constrained-plus zxy 0.5522 0.5534  
Somali Case-insensitive Constrained-plus TalTech 0.5690 0.5700  
Language Scoring Training Team WER CER Notes
Swahili Case-insensitive Constrained ustc_nelslip 0.3242 0.3259  
Swahili Case-insensitive Constrained zxy 0.3473 0.3489  
Swahili Case-insensitive Constrained TalTech 0.3498 0.3512  
Swahili Case-insensitive Constrained CRIM 0.3501 0.3513  
Swahili Case-insensitive Constrained THUEE 0.3524 0.3542  
Language Scoring Training Team WER CER Notes
Swahili Case-insensitive Constrained-plus THUEE 0.3149 0.3168  
Swahili Case-insensitive Constrained-plus zxy 0.3285 0.3307  
Swahili Case-insensitive Constrained-plus TalTech 0.3309 0.3324  
Language Scoring Training Team WER CER Notes
Swahili Case-sensitive Constrained zxy 0.4350 0.4353  
Swahili Case-sensitive Constrained THUEE 0.4371 0.4375  
Swahili Case-sensitive Constrained CRIM 0.4853 0.4857  
Language Scoring Training Team WER CER Notes
Swahili Case-sensitive Constrained-plus THUEE 0.4394 0.4399  
Swahili Case-sensitive Constrained-plus zxy 0.4399 0.4402  
Language Scoring Training Team WER CER Notes
Tagalog Case-insensitive Constrained ustc_nelslip 0.4038 0.4042  
Tagalog Case-insensitive Constrained THUEE 0.4170 0.4174  
Tagalog Case-insensitive Constrained TalTech 0.4186 0.4194  
Tagalog Case-insensitive Constrained CRIM 0.4318 0.4325  
Tagalog Case-insensitive Constrained zxy 0.4369 0.4375  
Tagalog Case-insensitive Constrained TranSpeech 0.7195 0.7200 No system description
Tagalog Case-insensitive Constrained Baymax 0.7445 0.7450 No system description
Tagalog Case-insensitive Constrained I2R 0.8366 0.8366  
Language Scoring Training Team WER CER Notes
Tagalog Case-insensitive Constrained-plus THUEE 0.3767 0.3772  
Tagalog Case-insensitive Constrained-plus TalTech 0.4017 0.4024  
Tagalog Case-insensitive Constrained-plus zxy 0.4132 0.4140  
Tagalog Case-insensitive Constrained-plus Baymax 0.7961 0.7963 No system description
Language Scoring Training Team WER CER Notes
Tagalog Case-sensitive Constrained zxy 0.4618 0.4617  
Tagalog Case-sensitive Constrained THUEE 0.4900 0.4896  
Tagalog Case-sensitive Constrained CRIM 0.5323 0.5322  
Language Scoring Training Team WER CER Notes
Tagalog Case-sensitive Constrained-plus zxy 0.4630 0.4629  
Tagalog Case-sensitive Constrained-plus THUEE 0.4641 0.4648  
Language Scoring Training Team WER CER Notes
Tamil Case-insensitive Constrained ustc_nelslip 0.6226 0.4130  
Tamil Case-insensitive Constrained CRIM 0.6382 0.4107  
Tamil Case-insensitive Constrained THUEE 0.6408 0.4482  
Tamil Case-insensitive Constrained TalTech 0.6423 0.4190  
Tamil Case-insensitive Constrained zxy 0.6577 0.4215  
Tamil Case-insensitive Constrained SRIB_ASR 0.8253 0.5787  
Language Scoring Training Team WER CER Notes
Tamil Case-insensitive Constrained-plus THUEE 0.5962 0.3810  
Tamil Case-insensitive Constrained-plus TalTech 0.6283 0.4055  
Tamil Case-insensitive Constrained-plus zxy 0.6323 0.4031  
Language Scoring Training Team WER CER Notes
Tamil Case-insensitive Unconstrained ustc_nelslip 0.5593 0.3859  
Tamil Case-insensitive Unconstrained zxy 0.5695 0.3966  
Language Scoring Training Team WER CER Notes
Vietnamese Case-insensitive Constrained ustc_nelslip 0.4031 0.3605  
Vietnamese Case-insensitive Constrained TalTech 0.4310 0.3899  
Vietnamese Case-insensitive Constrained zxy 0.4344 0.3879  
Vietnamese Case-insensitive Constrained CRIM 0.4398 0.3981  
Vietnamese Case-insensitive Constrained THUEE 0.4445 0.4016  
Vietnamese Case-insensitive Constrained I2R 0.7981 0.7404  
Language Scoring Training Team WER CER Notes
Vietnamese Case-insensitive Constrained-plus THUEE 0.3727 0.3127  
Vietnamese Case-insensitive Constrained-plus TalTech 0.4010 0.3610  
Vietnamese Case-insensitive Constrained-plus zxy 0.4044 0.3609  
Late submissions
Language Scoring Training Team WER CER Notes
Cantonese Case-insensitive Constrained CHNGA-DLMSL 0.6894 0.6331 Late, no system description
Language Scoring Training Team WER CER Notes
Cantonese Case-insensitive Constrained-plus CHNGA-DLMSL 0.6696 0.5848 Late, no system description
Language Scoring Training Team WER CER Notes
Farsi Case-insensitive Constrained CHNGA-DLMSL 0.9626 0.6339 Late, no system description
Language Scoring Training Team WER CER Notes
Kazakh Case-sensitive Constrained zxy 0.5106 0.3092 Late
Language Scoring Training Team WER CER Notes
Kurmanji-Kurdish Case-insensitive Constrained-plus THUEE 0.6019 0.5680 Late
Language Scoring Training Team WER CER Notes
Swahili Case-insensitive Unconstrained TranSpeech 0.4218 0.4239 Late, no Constrained submission, no system description
Language Scoring Training Team WER CER Notes
Tamil Case-insensitive Constrained THUEE 0.6391 0.4534 Late
Tamil Case-insensitive Constrained SRIB_ASR 0.7942 0.5772 Late
Tamil Case-insensitive Constrained I2R 1.0170 0.8558 Late
Language Scoring Training Team WER CER Notes
Vietnamese Case-insensitive Constrained-plus THUEE 0.3574 0.3148 Late

Table 2: OpenASR21 Results. Submissions with best WER for each participating team by language, scoring condition, and training condition.

System Descriptions

As part of the evaluation submission, teams were required to include a paper to describe their systems. Submitted system descriptions are provided below.

Teams were also encouraged to submit their updated work to be included in a Low-Resource ASR Development Special Session at INTERSPEECH 2022.

Disclaimer

NIST serves to coordinate the evaluations in order to support research and to help advance the state- of-the-art. NIST evaluations are not viewed as a competition, and such results reported by NIST are not to be construed, or represented, as endorsements of any team's system, or as official findings on the part of NIST or the U.S. Government.

Contact

Please email openasr_poc [at] nist.gov (openasr_poc[at]nist[dot]gov) for any questions or comments regarding the OpenASR Challenge.

Created March 17, 2022, Updated October 3, 2023