当前位置:网站首页>Google's AI model, which can translate 101 languages, is only one more than Facebook
Google's AI model, which can translate 101 languages, is only one more than Facebook
2020-11-08 12:56:00 【osc_1x6ycmfm】
Big data digest
source :VB
10 End of month ,Facebook Released a translatable 100 Machine learning models for languages , Microsoft has released a version that can translate 94 Models of languages , Google, of course, is not to be outdone .
Following Facebook After Microsoft , Google has open source a MT5 Model of , The model has achieved the most advanced results in a series of English natural language processing tasks .
MT5 It's Google's T5 Multilingual variants of the model , Is already containing 101 Pre training was conducted in data sets of six languages , Just like Facebook There's one more .
Github Address :
https://github.com/google-research/multilingual-t5
MT5 contain 3 Million to 130 One hundred million parameters , It can be directly applied to multiple language environments
MT5 contain 3 Million to 130 One hundred million parameters , It is reported that , It can learn 100 Multiple languages without interference .
MT5 Is in MC4 Trained on ,MC4 yes C4 A subset of ,MC4 Contains about 750GB The English text of , These texts come from Common Crawl The repository (Common Crawl Contains billions of web pages crawled from the Internet ). although C4 The dataset is explicitly designed to use only English , but MC4 covers 107 Languages , contain 10,000 Web pages or more .
however , There are still some deviations in the data set , Google researchers are trying to remove MC4 Duplicate lines in the document and filter pages with incorrect words to alleviate MT5 The deviation of . They also used tools to detect the main language of each page , And deleted credibility lower than 70% The page of .
Google said , maximal MT5 The model has 130 One hundred million parameters , More than the 2020 year 10 All benchmarks for monthly testing . Of course , Whether the benchmark fully reflects the real performance of the model , This is a topic worthy of debate .
Some research shows that , Open Domain Question Answering Model (Open-Domain Question-Answering, A model that can theoretically answer novel questions with novel answers ) It's usually just a matter of simply remembering the answers found in the training data based on the data set . But Google researchers assert that MT5 It's a step towards a powerful model , These functions do not require challenging modeling techniques .
Google researchers describe MT5 In his paper, he wrote ,“ in general , Our findings highlight the importance of model competence in cross language representation learning , And show , By relying on filtering 、 Parallel data or intermediate tasks , Expanding the simple pre training formula is a viable alternative .”“ We demonstrated T5 Recipes are directly applicable to Multilingual Settings , And it achieves powerful performance on different benchmark sets .”
comparison Facebook And Microsoft , Google's MT5 It seems to be a little better
Facebook The new model of is called M2M-100,Facebook Claim to be the First Multilingual Machine Translation Model , Can be directly in 100 Translate back and forth between any pair of languages .Facebook AI Build a total of 100 Language 75 A huge data set of hundreds of millions of sentences . Using this dataset , The research team trained a man with more than 150 A universal translation model with hundreds of millions of parameters , According to the Facebook A blog description of , The model can “ Get information about the relevant language , And reflect a more diverse language text and language form ”.
And Microsoft's machine learning translation model is called T-ULRv2, It can be translated. 94 Languages . Microsoft claims ,T-ULRv2 stay XTREME( A natural language processing benchmark created by Google ) Got the best search results in , And will use it to improve Word Semantic search in 、Outlook and team Reply suggestions and other functions in .
T-ULRv2 stay XTREME At the top of the list
T-ULRv2 It is a joint research product of Microsoft Research Institute and Turing team , contain 5.5 One hundred million parameters , The model uses these parameters to predict . Microsoft researchers trained on a multilingual data corpus T-ULRv2, The data corpus comes freely 94 Web pages made up of languages . In the process of training ,T-ULRv2 Translation by predicting the hidden words in sentences of different languages , Occasionally, contextual cues are obtained from paired translations of English and French .
All in all , In terms of the number of languages translated , Google's MT5 It seems to be a little better , But large numbers don't mean high accuracy , Just Google and Facebook For two translation models , There is still room for improvement in the translation of some low resource languages , Like wolov 、 Malathi . Besides , Each machine learning model will have a certain deviation , Just like Allen AI What the researchers at the Institute said ,“ The existing machine learning technology can not avoid this defect , People are in urgent need of better training mode and model construction ”.
Relevant reports :
https://venturebeat.com/2020/10/26/google-open-sources-mt5-a-multilingual-model-trained-on-over-101-languages/
https://venturebeat.com/2020/10/20/microsoft-details-t-urlv2-model-that-can-translate-between-94-languages/
The main work of the future intelligent laboratory includes : establish AI Intelligence system intelligence evaluation system , Carry out the world artificial intelligence IQ evaluation ; Launch the Internet ( City ) Cloud brain research project , Building the Internet ( City ) Cloud brain technology and enterprise map , For the promotion of enterprises , Intelligent level service of industry and city .
If you are interested in laboratory research , Welcome to the future intelligent laboratory online platform . Scan the QR code below or click on the bottom left corner of this article “ Read the original ”
版权声明
本文为[osc_1x6ycmfm]所创,转载请带上原文链接,感谢
边栏推荐
- Ali! Visual computing developer's series of manuals (with internet disk link)
- Powershell 使用.Net对象发送邮件
- 【Python 1-6】Python教程之——数字
- 一文读懂机器学习“数据中毒”
- We interviewed the product manager of SQL server of Alibaba cloud database, and he said that it is enough to understand these four problems
- 2018中国云厂商TOP5:阿里云、腾讯云、AWS、电信、联通 ...
- 来自朋友最近阿里、腾讯、美团等P7级Python开发岗位面试题
- PDMS cutting software
- Istio traffic management -- progress gateway
- Iqkeyboardmanager source code to see
猜你喜欢
python基本语法 变量
Flink从入门到真香(10、Sink数据输出-Elasticsearch)
Tight supply! Apple's iPhone 12 power chip capacity exposed
应届生年薪35w+ !倒挂老员工,互联网大厂薪资为何越来越高?
数据库连接报错之IO异常(The Network Adapter could not establish the connection)
Windows下快递投递柜、寄存柜的软件初探
PMP experience sharing
华为云重大变革:Cloud&AI 升至华为第四大 BG ,火力全开
Flink从入门到真香(7、Sink数据输出-文件)
Top 5 Chinese cloud manufacturers in 2018: Alibaba cloud, Tencent cloud, AWS, telecom, Unicom
随机推荐
It's worth seeing! EMR elastic low cost offline big data analysis best practice (with network disk link)
新的目标市场在哪里?锚定的产品是什么?| 十问2021中国企业服务
[Python 1-6] Python tutorial 1 -- number
C language I blog assignment 03
This paper analyzes the top ten Internet of things applications in 2020!
Written interview topic: looking for the lost pig
笔试面试题目:求丢失的猪
Implementation of verification code recognition in Python opencv pytesseract
阿里撕下电商标签
Understanding design patterns
Research on WLAN direct connection (peer-to-peer connection or P2P) and cross platform research of IOS
2018中国云厂商TOP5:阿里云、腾讯云、AWS、电信、联通 ...
笔试面试题目:判断单链表是否有环
Don't look! Full interpretation of Alibaba cloud's original data lake system! (Internet disk link attached)
Enabling education innovation and reconstruction with science and technology Huawei implements education informatization
笔试面试题目:盛水最多的容器
What is SVG?
Tidb performance competition 11.02-11.06
Where is the new target market? What is the anchored product? |Ten questions 2021 Chinese enterprise service
The young generation of winner's programming life, the starting point of changing the world is hidden around