Modern universal language models and transformers such as BERT, ELMo, XLNet, RoBERTa and others need to be properly compared and evaluated. In the last year, new models and methods for pretraining and transfer learning have driven striking performance improvements across a range of language understanding tasks.
We offer testing methodology based on tasks, typically proposed for “strong AI” — logic, commonsense, reasoning. Adhering to the GLUE and SuperGLUE methodology, we present a set of test tasks for general language understanding and leaderboard models.
For the first time a complete test for Russian language was developed, which is similar to its English analog. Many datasets were composed for the first time, and a leaderboard of models for the Russian language with comparable results is also presented.
@article{shavrina2020russiansuperglue,
title={RussianSuperGLUE: A Russian Language Understanding Evaluation Benchmark},
author={Shavrina, Tatiana and Fenogenova, Alena and Emelyanov, Anton and Shevelev, Denis and Artemova, Ekaterina and Malykh, Valentin and Mikhailov, Vladislav and Tikhonova, Maria and Chertok, Andrey and Evlampiev, Andrey},
journal={arXiv preprint arXiv:2010.15925},
year={2020}
}
rybolos@gmail.com
Tatiana Shavrina is a PhD student at the Higher School of Economics and is the Head of RnD Department in NLP at Sberbank. Tatiana has presented her main works in the field of web сorpus сonstruction and universal transformers. Tatiana has also organized evaluation tracks for Russian in spelling correction, morphology, full UD annotation, ellypsis resolution.
alenush93@gmail.com
Alena Fenogenova has a Master degree in Computational linguistics at the Higher School of Economics, Moscow. She has a number of publications in NLP field. Now Alena works in Sberbank in NLP research team. Her research interests are language understanding, Question Answering, Argument Mining, etc.
valentin.malykh@huawei.com
Valentin Malykh has written his PhD thesis at Moscow Institute of Physics and Technology and defended it at Institute for Systems Programming, Russian Academy of Sciences in 2019. Dr. Malykh has more than 20 papers in NLP field, including publications on such conferences as NeurIPS, ACL, WSDM. Now Valentin is employed as a senior research scientist at Huawei Noah`s Ark laboratory.
login-const@mail.ru
Anton Emelyanov writes his PhD thesis at Moscow Institute of Physics and Technology. Now He is a member of the Laboratory for Computational Pragmatics. Anton is also employed as a research scientist at Sberbank, RnD in NLP team. His interests are language understanding, language models and generation.
m_tikhonova94@mail.ru
Maria Tikhonova is a PhD student in Computer Science at the Higher School of Economics, Moscow. She received a Masters’s degree in fundamental mechanics and mathematics at the Moscow State University. She works in Sberbank in the Research&Development in NLP department. Her interests are NLP and topic modeling.
reddraner@gmail.com
Denis Shevelev is a graduate of the Sholokhov State University (Moscow) and has two higher educations in the field of philology and journalism. Since 2005, works in copywriting and editing. His interests are NLU, sentiment analysis, and general intelligence. He is also the author of two graphic novels.
glushkovato@gmail.com
Research intern at Higher School of Economics / Moscow, Russia
glushkovato@gmail.com
Taisiya has a Master degree in Computer Science at the Higher School of Economics, where she currently holds a Research Intern position in the research group in natural language processing. Previously worked at Unbabel, Diginetica and Sberbank.