当前位置:网站首页>ByteDance open source Gan model compression framework, saving up to 97.8% of computing power - iccv 2021
ByteDance open source Gan model compression framework, saving up to 97.8% of computing power - iccv 2021
2022-07-06 16:49:00 【ByteDance Technology】
Put the outline of the shoes :
Restore to the shoe body :
How much calculation is needed ?
Use the most basic Pix2Pix Model , Need to consume 56.8G;
And when Pix2Pix The model is compressed by a new technology , It only needs 1.219G, It's the original 1/46.6, It greatly saves the amount of calculation .
The technology used here , That is, the technical team of ByteDance will meet at the top of computer vision ICCV 2021 Published on Online multi particle Distillation Algorithm (Online Multi-Granularity Distillation, abbreviation OMGD).
This model compression framework , To govern GAN The volume of the model is too large 、 Too much calculation effort , At present, the code is open source ( The address is at the end of the article ), as well as CycleGAN And Pix2Pix Pre training model of , And it has been landing on products such as Tiktok .
Compared with similar model compression algorithms ,OMGD This new framework , Not only the pressure is smaller , And the pressure is better .
For example, in the process of turning horses into zebras :
MACs Indicates the amount of calculation consumed , The number in brackets is the promotion multiple
And turning summer into winter :
Restore the divided street view into photos ( Pay attention to the cyclist ):
Experiments show that , This technology can make GAN The calculation amount of the model is reduced to the original 1/46、 The minimum number of parameters is reduced to the original 1/82.
Convert it , Is to omit 97.8% Amount of computation .
OMGD How did it happen
Model compression usually uses 「 Distillation of knowledge 」 Methods , That is, there are many parameters 、 The bulky model acts as 「 Teacher model 」, There are few parameters to supervise and optimize 、 Small size 「 Student model 」, Let the student model get the knowledge taught by the teacher model without expanding the volume .
OMGD This technology optimizes a student model by two teacher models that complement each other in the width and depth of the neural network , whole Pipeline That's true :
The framework transfers the concepts of different levels from the middle layer and the output layer , It can be used without discriminator and without Ground Truth Training under the setting of , Realize the refinement of knowledge , The on-line distillation scheme was optimized as a whole .
stay Pix2Pix and CycleGAN Experimental data on two well-known models show ,OMGD You can use the least number of parameters 、 The lowest amount of computation achieves the best image generation effect .
The most the right side FID The smaller the score , Indicates that the better the generation effect
Why make the big model smaller ?
ByteDance technical team related research and development students said , This is the first online knowledge distillation to compress GAN Technical solution , Has landed in Tiktok .
you 're right , All kinds of funny special effects props you see in Tiktok , All kinds of algorithm models are needed to run , Especially in dealing with special effects related to images ,GAN It's a general method , This set of GAN The model compression scheme has also been implemented in Tiktok , such as 「 Motion pictures 」:
「 Dance gloves 」:
also 「 Three screen life 」:
however ,GAN Models are usually large , It takes a lot of calculation to complete , It is a great challenge to land on mobile phones, especially low-end computers with insufficient computing power .OMGD A classmate of the R & D team said :“ We will measure the coverage of the model , That is, how many models of a model can run smoothly , After successfully compressing the model, more mobile phones can be covered , Let more people use , If the original model may be iPhone 11 Ability to use , After the compression iPhone 7 Can also be used .”
therefore , Model compression is a rigid requirement , How to make GAN Used by more people 、 Provide more inclusive services , It is the direction that the technology industry has been pursuing .
ByteDance technology team was the first 2017 Research on compression of annual input model , Today's headline 、 Tiktok 、 Cut and reflect 、 Watermelon video, etc App All have relevant technologies , I've also got 2020 year IEEE Low power computer vision challenge (LPCV) The champion of two tracks .
after OMGD Before compression , Teams usually use distillation or pruning algorithms to complete GAN Compression of models , Because the input resolution required by the model is very large , The amount of calculation is still very large , Not compressed to the extreme .
How can we achieve more extreme compression ?
After studying a large number of existing methods in academia , ByteDance technical team students did not find a method suitable for the company's business , Instead, I decided to study by myself , Creatively in GAN In model compression, it is the first time to think of using two complementary teacher models to train the same student model , And achieved the success of the experiment .
Now? ,OMGD In practice, it can be faster than the original method 20~30%, Some can even reach 80%.
also , As can be 「 Online compression 」 Methods ,OMGD Greatly reduced GAN The complexity of model production . there 「 On-line 」 It's not the online state that we refer to in our daily life , It means that the distillation process is completed in one step ,“ Previous GAN The compression method is carried out in several steps , For example, pre training first , Train again after compression , Then there are other steps , It's more complicated as a whole ; Our new method can complete the whole process in one step , The effect is also much better than other methods .” The team technical classmate said .
Now? , This kind of model compression technology can not only save computing power and energy , It can provide users with a smooth experience , Help creators inspire creativity , Enrich your life .
Related links
Address of thesis :
https://arxiv.org/abs/2108.06908
GitHub Code and training model :
https://github.com/bytedance/OMGD
边栏推荐
- 字节跳动技术新人培训全记录:校招萌新成长指南
- 第5章 NameNode和SecondaryNameNode
- ~71 abbreviation attribute of font
- LeetCode 1638. Count the number of substrings with only one character difference
- ~84 form supplement
- LeetCode 1641. Count the number of Lexicographic vowel strings
- Research Report on market supply and demand and strategy of Chinese table lamp industry
- Codeforces Global Round 19
- 解决Intel12代酷睿CPU单线程调度问题(二)
- LeetCode 1640. Can I connect to form an array
猜你喜欢
Usage of insert() in vector
Hbuilder x format shortcut key settings
Two weeks' experience of intermediate software designer in the crash soft exam
Record the error reason: terminate called after throwing an instance
ffmpeg命令行使用
【锟斤拷】的故事:谈谈汉字编码和常用字符集
Spark独立集群Worker和Executor的概念
Use JQ to realize the reverse selection of all and no selection at all - Feng Hao's blog
第5章 NameNode和SecondaryNameNode
图像处理一百题(11-20)
随机推荐
~73 other text styles
~86m rabbit practice
Li Kou leetcode 280 weekly match
Market trend report, technical innovation and market forecast of tabletop dishwashers in China
(multiple methods, need to continue to see) 7-11 go deep into the tiger's Den
Chapter 7__ consumer_ offsets topic
【锟斤拷】的故事:谈谈汉字编码和常用字符集
Hbuilder X格式化快捷键设置
Two weeks' experience of intermediate software designer in the crash soft exam
解决Intel12代酷睿CPU单线程调度问题(二)
Basic principles of video compression coding and audio compression coding
Usage of insert() in vector
Research Report on market supply and demand and strategy of double drum magnetic separator industry in China
字节跳动多篇论文入选 CVPR 2021,精选干货都在这里了
音视频开发面试题
业务系统从Oracle迁移到openGauss数据库的简单记录
LeetCode 1557. The minimum number of points that can reach all points
Chapter 5 namenode and secondarynamenode
FLV格式详解
简单尝试DeepFaceLab(DeepFake)的新AMP模型