Home / Research / Academics / Content
The “Vega” of Wuhan University wins world championship
Author:Jiang Zhaoxi  Date:2022-06-02  Clicks:

The Vega v1 model proposed by the Wuhan University - JD Trusted Artificial Intelligence Joint Research Center has topped the General Language Understanding Evaluation (GLUE) list, a prestigious global natural language processing list,  with an overall average score of 91.3, setting a new world record in this area.

Vega” makes its stunning debut

The GLUE list, jointly launched by New York University, the University of Washington, Google AI offshoot DeepMind and other institutions, is considered an important metric to measure natural language processing and pre-training techniques.

In the recently released  GLUE list, Vega v1, a natural language processing model with a very large parameter scale proposed by Wuhan University - JD Trusted Artificial Intelligence Joint Research Center, scored 91.3 and beat Microsoft, Facebook, and Stanford University, clearly demonstrating Vega v1’s leading position in artificial intelligence technology.

GLUE list ranking chart

This mysterious and romantic name comes from α Lyrae, the super large-scale computing cluster of JD Explore Academy . It is with its support that such large-scale training can be realized. Vega is an alias for α Lyrae, the brightest star in the constellation Lyra. The team wanted the Vega v1 model to be the most special among the pre-training models.

Team members working on the computer

As a general-purpose language model, the Vega v1 model can be applied to a variety of natural language processing tasks and has a wide range of future applications, such as intelligent question and answer functionality, chatbot, grammar correction, and autonomous driving. By adopting model compression, pruning and distillation to lighten Vega v1, a model with smaller parameters can be obtained and deployed in an intelligent terminal, making peoples daily life more convenient.

In addition to the powerful capabilities of the model itself, the team has also adopted many matching and fine-tuning strategies to efficiently  update the model’s parameters with a small number of annotated samples for specific downstream natural language processing tasks, effectively improving the accuracy of the Vega v1 model.

Its a breakthrough to  new heights

A lack of generalization is ubiquitous in artificial intelligence. For instance, for each AI task, it is often necessary to train a specific model using a relevant data set. The same specific model that performs well with the current task may not perform as well on  other tasks.

To address this challenge and broaden the generality of artificial intelligence, more and more AI  employs generic pre-training model. Good results can be achieved by training a generic model on a large-scale data set and fine-tuning it for a specific task,  effectively solving the problem of insufficient generalizability of models.

The Vega v1 model, as a large-scale pre-training language model, has also achieved good results on a variety of downstream tasks. Compared with other models on the GLUE list, Vega v1 achieves breakthroughs in several pre-training techniques such as an  energy-efficient parallelized training framework and data utilization approach, the innovative model architecture with billions of parameters, an improved self-supervised pre-training goal, and allowing the model to learn whole-sentence representations based on different granularities of words, phrases, and short sentences, enabling multi-granular sentence-level representations. These all make the model itself more competitive.

The test results of the Vega v1 model

The GLUE list covers a total of nine major NLP tasks such as natural language inference, semantic textual similarity, question and answer. The human test results for each task are provided at the initial stage  of the list’s establishment, representing the level of human intelligence for each task. With the continuous research into pre-training models, such models have been able to outperform human test results on several tasks in GLUE, but sentiment analysis and co-reference resolution have performed more poorly than humans.

The Vega v1 model not only topped the GLUE list with the first overall average score, but also surpassed the human test results on these two challenging tasks for the first time, indicating that it has taken the intelligence of pre-training models to a new level.

In the future, the team will also consider incorporating technologies such as trusted AI to fully upgrade the Vega v1 model by integrating trusted artificial intelligence and other technologies to enhance interpretability, privacy and fairness while constantly improving text understanding capabilities.

The dream makers behind Vega v1

As a research institute jointly established by the WHU Artificial Intelligence Institute, the School of Computer Science and JD, the Wuhan University - JD Trusted Artificial Intelligence Joint Research Center has published dozens of high-level research papers since its establishment in 2021. Additionally, it won first place both in the Video + depth branch of the ICCV-2021 Benchmarking Multi-Target Tracking Competitions and the GLUE Competition, the top test in global natural language processing.

Group photo of the Vega v1 model R&D team

In the process of model training and competition, the team also encountered many difficulties. Due to the lack of experience in large-scale model training, they had to learn many things from scratch; the heavy demand on computing resources for model training poses severe challenges for the effective management of these resources. In the face of these difficulties, the team worked together to analyze the problems, debug codes, and discuss solutions until the early morning. It is these efforts that allow the Vega v1 model to be constantly optimized and improved.

Dr. Zhong Qihuang, the core member of the team, believes that both learning and research require composure and dedication. Choose a direction, set a time, and the rest is just hard work and persistence. Time will give us the final answer. In this way, they have been able to overcome difficulties and achieve excellence in the field of artificial intelligence research, just like the star Vega, which shines brightly in the sky.


Like Vega shining in the sky,

The Vega v1 model also yields brilliant results

In its own domain.

On the path of pursuing dreams,

May we all keep shining

In the company of stars.


Rewritten by Zhou Chuangyu

Edited by Su Xinyue, Zou Xiaohan, Sylvia, Xi Bingqing


Prev Section:Academician Li Deren becomes China’s first recipient of the Brock Gold Medal Award
Next Section:Application of new photosensitive protein brings hope to 20 million blind people to regain vision

[ Close ]

Copyright @ 2014 Wuhan University