当前位置:网站首页>AI super clear repair resurfaces the light in Huang Jiaju's eyes, Lecun boss's "deep learning" course survival report, beautiful paintings only need one line of code, AI's latest paper | showmeai info

AI super clear repair resurfaces the light in Huang Jiaju's eyes, Lecun boss's "deep learning" course survival report, beautiful paintings only need one line of code, AI's latest paper | showmeai info

2022-07-07 00:33:00 ShowMeAI

ShowMeAI daily Series new upgrade ! Cover AI Artificial intelligence Tools & frame | project & Code | post & Share | data & resources | Research & The paper Wait for the direction . Click to see List of historical articles , Subscribe to topics in the official account #ShowMeAI Information daily , Can receive the latest daily push . Click on Thematic collection & Electronic monthly Quickly browse the complete works of each topic . Click on here Reply key daily Free access AI Electronic monthly and information package .

AI Repair huangjiaju 31 Classic pictures years ago

7 month 3 Friday night , utilize AI Technology repair Beyond『1991 Life contact concert 』, Time interval 31 It was re shown on Tiktok platform in . This super clear repair remake , Insufficient resolution 540p To be close to 4K, The skin 、 The smile and other details are all clearly visible .

Tools & frame

DiscoArt】 Use one line of code Disco Diffusion Generate artwork

GitHub: https://github.com/jina-ai/discoart

Disco Diffusion Is a Google Colab Notebook Free tools , Use Python Language creates beautiful images and videos from text .DiscoArt Simplify the process , Make the results reproducible 、 persistence 、gRPC/HTTP service 、 Later analysis is more concise , Help artists and AI Lovers create more gracefully .

Versatile Data Kit (VDK) 】 Use SQL or Python Knowledge builds data pipeline

GitHub: https://github.com/vmware/versatile-data-kit

VDK( Multifunctional data toolkit ) It's an open source framework , from Data SDK and Control Service Two kits . Data engineers can use both SQL or Python Language , Development 、 Deploy 、 Run and manage data jobs . The following figure shows the data flow and VDK The role of .

The Nettle Magic Project】 Poker barcode scanner , Determine the position of each card by scanning the QR code composed of the side of the poker

GitHub: https://github.com/nettlep/magichttps://github.com/nettlep/magic

Find ink that can only be recognized under specific infrared conditions , And mark the edge of a deck of playing cards . Use a NoIR Raspberry pie for camera Zero W These markers can be detected , coordination iPad Running on iOS Client applications , You can clearly identify playing cards and infrared marks .

wenet_trt8】 use TRT8 Deploy the open source speech recognition toolkit

GitHub: https://github.com/huismiling/wenet_trt8

WeNet It is a speech recognition toolkit for industrial applications , Model from encoder and decoder Two parts . This project uses FP32 Model (Pytorch Directly derived ), Use ORT(ONNX Run Time) and PPQ Two quantitative methods are right WeNet The model is quantified .

  • Use ORT quantitative : Model test results WER:6.06%, Time consuming :Encoder 7.34ms,Decoder 3.32ms. Data set effect decreases 1.46%, and Encoder Speed up 2.5 times ,Decoder Speed up 6.3 times .
  • Use PPQ quantitative : Model test results WER:5.71%, Time consuming :Encoder 12.14ms,Decoder 2.68ms. Data set effect decreases 1.11%, and Encoder Speed up 1.5 times ,Decoder Speed up 7.7 times .

post & Share

New York University 《 Deep learning (2021) 》 Course summary and study notes

note : https://dimgeo.com/posts/NYU-2021-DL-course-overview-and-notes/

Course : https://atcold.github.io/NYU-DLSP21/

NYU 2021《DEEP LEARNING》 Lecturer of , One is famous Yann LeCun, The other is an assistant professor in the Department of computer science at New York University Alfredo. Course about 50 Hours , Information is very dense 、 It covers a wide range of 、 And it includes Facebook AI The guest lecturer of the research laboratory showed CV and NLP The latest development of the industry .

however ! Be prepared !LeCun Our courses are not like Andrew NG Wu Enda is so friendly , And you need to have basic machine learning knowledge and linear algebra knowledge .

Markov chain for programmers 》 Free ebooks

Github: https://github.com/czekster/markov

pdf: https://czekster.github.io/markov/MC-for-programmers2022.pdf

Author of this book Ricardo M. Czekster, Explain Markov chain and its basic solution , contain Markov Chains、DTMC、CTMC Other chapters ( See the picture below for the table of contents ). Can be in GitHub Find the basic materials of this book on the page , for example C Programming code and solutions 、MATLAB Script 、 This book provides examples of prismatic models (CTMC/DTMC) etc. .

2022 China's future Unicorn TOP100 The list

At the 6th conference on the growth of all things · China future Unicorn Conference ,《2022 China's future Unicorn TOP100 The list 》 Official release , Focus on enterprise services 、 Artificial intelligence 、 The new consumer 、 Advanced manufacturing 、 Meta universe 5 Large area , And selected in various fields 20 High quality enterprises . The list is here ! You my AI Little cute people reference ~

data & resources

University of Geneva, Switzerland 《 Deep learning 》 Course videos and materials

Address : https://fleuret.org/dlc/

UNIGE 14x050《Deep Learning》 The course provides a comprehensive introduction to in-depth learning , With PyTorch Examples in the framework :

  • Machine learning goals and major challenges
  • Tensor operation
  • Automatic differentiation , gradient descent
  • Deep learning of specific technologies
  • Generated 、 Cyclic 、 Attention model

Super resolution / hallucination Related work list

GitHub: https://github.com/junjun-jiang/Face-Hallucination-Benchmark

Face-Hallucination-Benchmark:A collection of state-of-the-art face super-resolution/hallucination methods by Junjun Jiang

Research & The paper

You can click on the here Reply key daily , Get the collected papers for free .

Official account reply daily , Get the collected papers for free .

  • Research progress

  • 2022 year 6 month 27 Japan 【 Deep learning 】| Benchopt: Reproducible, efficient and collaborative optimization benchmarks

  • ICLR 2020【 Figure neural network 】| Measuring and Improving the Use of Graph Information in Graph Neural Networks

  • 2022 year 6 month 27 Japan 【 natural language processing 】| Center-Embedding and Constituency in the Brain and a New Characterization of Context-Free Languages

  • 2022 year 6 month 27 Japan 【 natural language processing 】| Endowing Language Models with Multimodal Knowledge Graph Representations

  • 2022 year 6 month 27 Japan 【 machine learning 】| Fast sequence to graph alignment using the graph wavefront algorithm

  • 2022 year 6 month 27 Japan 【 Computer vision 】| CLTS-GAN: Color-Lighting-Texture-Specular Reflection Augmentation for Colonoscopy

The paper :Benchopt: Reproducible, efficient and collaborative optimization benchmarks

Paper title :Benchopt: Reproducible, efficient and collaborative optimization benchmarks

Paper time :27 Jun 2022

Field Deep learning

Corresponding tasks :Image Classification,Stochastic Optimization, Image classification , Stochastic optimization

Address of thesis https://arxiv.org/abs/2206.13424

Code implementation https://github.com/benchopt/benchopt

Author of the paper :Thomas Moreau, Mathurin Massias, Alexandre Gramfort, Pierre Ablin, Pierre-Antoine Bannier, Benjamin Charlier, Mathieu Dagréou, Tom Dupré La Tour, Ghislain Durif, Cassio F. Dantas, Quentin Klopfenstein, Johan Larsson, En Lai, Tanguy Lefort, Benoit Malézieux, Badr Moufad, Binh T. Nguyen, Alain Rakotomamonjy, Zaccharie Ramzi, Joseph Salmon, Samuel Vaiter

Brief introduction of the paper :Numerical validation is at the core of machine learning research as it allows to assess the actual impact of new methods, and to confirm the agreement between theory and practice./ Numerical verification is the core of machine learning research , Because it can evaluate the actual impact of the new method , And confirm the consistency between theory and practice .

Abstract of paper :Numerical validation is at the core of machine learning research as it allows to assess the actual impact of new methods, and to confirm the agreement between theory and practice. Yet, the rapid development of the field poses several challenges: researchers are confronted with a profusion of methods to compare, limited transparency and consensus on best practices, as well as tedious re-implementation work. As a result, validation is often very partial, which can lead to wrong conclusions that slow down the progress of research. We propose Benchopt, a collaborative framework to automate, reproduce and publish optimization benchmarks in machine learning across programming languages and hardware architectures. Benchopt simplifies benchmarking for the community by providing an off-the-shelf tool for running, sharing and extending experiments. To demonstrate its broad usability, we showcase benchmarks on three standard learning tasks: ℓ2-regularized logistic regression, Lasso, and ResNet18 training for image classification. These benchmarks highlight key practical findings that give a more nuanced view of the state-of-the-art for these problems, showing that for practical evaluation, the devil is in the details. We hope that Benchopt will foster collaborative work in the community hence improving the reproducibility of research findings.

Numerical verification is the core of machine learning research , Because it can evaluate the actual impact of the new method , And confirm the consistency between theory and practice . However , The rapid development of this field has brought some challenges : Researchers face a large number of methods that need to be compared 、 Limited transparency and consensus on best practices , And tedious re implementation . therefore , Verification is often very one-sided , This may lead to wrong conclusions , Slow down the progress of research . We proposed Benchopt, This is a cooperation framework , For Automation 、 Replicate and publish optimization benchmarks for machine learning across programming languages and hardware architectures .Benchopt By providing an off the shelf tool to run 、 Share and expand experiments , This simplifies the benchmarking of the community . To prove its wide availability , We show benchmarks for three standard learning tasks :ℓ2-regularized logistic regression, Lasso, and ResNet18 Image classification training . These benchmarks highlight key practical findings , Give a more detailed view of the most advanced level of these issues , It shows that for the actual evaluation , The key lies in the details . We hope Benchopt It can promote community cooperation , So as to improve the repeatability of the research results .

The paper :Measuring and Improving the Use of Graph Information in Graph Neural Networks

Paper title :Measuring and Improving the Use of Graph Information in Graph Neural Networks

Paper time :ICLR 2020

Field Figure neural network

Corresponding tasks :Representation Learning, Representational learning

Address of thesis https://arxiv.org/abs/2206.13170

Code implementation https://github.com/yifan-h/CS-GNN

Author of the paper :Yifan Hou, Jian Zhang, James Cheng, Kaili Ma, Richard T. B. Ma, Hongzhi Chen, Ming-Chang Yang

Brief introduction of the paper :Graph neural networks (GNNs) have been widely used for representation learning on graph data./ Figure neural network (GNNs) It has been widely used in graph data representation learning .

Abstract of paper :Graph neural networks (GNNs) have been widely used for representation learning on graph data. However, there is limited understanding on how much performance GNNs actually gain from graph data. This paper introduces a context-surrounding GNN framework and proposes two smoothness metrics to measure the quantity and quality of information obtained from graph data. A new GNN model, called CS-GNN, is then designed to improve the use of graph information based on the smoothness values of a graph. CS-GNN is shown to achieve better performance than existing methods in different types of real graphs.

Figure neural network (GNNs) It has been widely used in graph data representation learning . However , People are right. GNNs There is limited understanding of how much performance is actually obtained from the graph data . This paper introduces a context based GNN frame , Two smoothness indexes are proposed to measure the quantity and quality of information obtained from graph data . Then I designed a new GNN Model , be called CS-GNN, To improve the use of graph information based on graph smoothness values .CS-GNN It has been proved that it has better performance than the existing methods in different types of real graphs .

The paper :Center-Embedding and Constituency in the Brain and a New Characterization of Context-Free Languages

Paper title :Center-Embedding and Constituency in the Brain and a New Characterization of Context-Free Languages

Paper time :27 Jun 2022

Field natural language processing

Corresponding tasks :Dependency Parsing, Dependency resolution

Address of thesis https://arxiv.org/abs/2206.13217

Code implementation https://github.com/dmitropolsky/assemblies

Author of the paper :Daniel Mitropolsky, Adiba Ejaz, Mirah Shi, Mihalis Yannakakis, Christos H. Papadimitriou

Brief introduction of the paper :We address two of the most important questions left open by that work: constituency (the identification of key parts of the sentence such as the verb phrase) and the processing of dependent sentences, especially center-embedded ones./ We discussed the two most important issues left by this work : Word formation ( Identify the key parts of the sentence , Such as verb phrases ) And subordinate sentence processing , Especially the sentence embedded in the center .

Abstract of paper :A computational system implemented exclusively through the spiking of neurons was recently shown capable of syntax, that is, of carrying out the dependency parsing of simple English sentences. We address two of the most important questions left open by that work: constituency (the identification of key parts of the sentence such as the verb phrase) and the processing of dependent sentences, especially center-embedded ones. We show that these two aspects of language can also be implemented by neurons and synapses in a way that is compatible with what is known, or widely believed, about the structure and function of the language organ. Surprisingly, the way we implement center embedding points to a new characterization of context-free languages.

lately , A computing system implemented entirely by neurons has been proved to be able to perform syntax , That is, dependency analysis of simple English sentences . We discussed the two most important issues left by this work : Word formation ( Identify the key parts of the sentence , Such as verb phrases ) And subordinate sentence processing , Especially the sentence embedded in the center . We show that , These two aspects of language can also be realized by neurons and synapses , Its way is consistent with the known or widely recognized structure and function of language organs . It's amazing , The way we implement central embedding points to the new features of contextless language .

The paper :Endowing Language Models with Multimodal Knowledge Graph Representations

Paper title :Endowing Language Models with Multimodal Knowledge Graph Representations

Paper time :27 Jun 2022

Field natural language processing

Corresponding tasks :Multilingual Named Entity Recognition,named-entity-recognition,+1

Address of thesis https://arxiv.org/abs/2206.13163

Code implementation https://github.com/iacercalixto/visualsem

Author of the paper :Ningyuan Huang, Yash R. Deshpande, Yibo Liu, Houda Alberts, Kyunghyun Cho, Clara Vania, Iacer Calixto

Brief introduction of the paper :We use the recently released VisualSem KG as our external knowledge repository, which covers a subset of Wikipedia and WordNet entities, and compare a mix of tuple-based and graph-based algorithms to learn entity and relation representations that are grounded on the KG multimodal information./ We use the recently released VisualSem KG As our external knowledge base , It covers Wikipedia and WordNet A subset of entities , And compares the hybrid algorithm based on tuple and graph to learn based on KG Entity and relationship representation of multimodal information .

Abstract of paper :We propose a method to make natural language understanding models more parameter efficient by storing knowledge in an external knowledge graph (KG) and retrieving from this KG using a dense index. Given (possibly multilingual) downstream task data, e.g., sentences in German, we retrieve entities from the KG and use their multimodal representations to improve downstream task performance. We use the recently released VisualSem KG as our external knowledge repository, which covers a subset of Wikipedia and WordNet entities, and compare a mix of tuple-based and graph-based algorithms to learn entity and relation representations that are grounded on the KG multimodal information. We demonstrate the usefulness of the learned entity representations on two downstream tasks, and show improved performance on the multilingual named entity recognition task by 0.3%–0.7% F1, while we achieve up to 2.5% improvement in accuracy on the visual sense disambiguation task. All our code and data are available in https://github.com/iacercalixto/visualsem-kg

We propose a way , Through the external knowledge map (KG) Store knowledge in and use dense indexes from this KG To retrieve the , Make the parameter efficiency of natural language understanding model higher . Given the downstream task data ( It may be multilingual ), For example, German sentences , We from KG Retrieve entities in , And use their multimodal representations to improve the performance of downstream tasks . We use the recently released VisualSem KG As our external knowledge base , It covers Wikipedia and WordNet A subset of entities , And compares the hybrid algorithm based on tuple and graph to learn based on KG Entity and relationship representation of multimodal information . We proved the usefulness of the learned entity representation in two downstream tasks , It also shows that the performance in multilingual named entity recognition task is improved 0.3%-0.7% F1, In the task of ambiguity in the visual sense , We achieved up to 2.5% Improve the accuracy of . All our codes and data can be found on the following website https://github.com/iacercalixto/visualsem-kg

The paper :Fast sequence to graph alignment using the graph wavefront algorithm

Paper title :Fast sequence to graph alignment using the graph wavefront algorithm

Paper time :27 Jun 2022

Field machine learning

Address of thesis https://arxiv.org/abs/2206.13574

Code implementation https://github.com/lh3/gwfa

Author of the paper :Haowen Zhang, Shiqi Wu, Srinivas Aluru, Heng Li

Brief introduction of the paper :Motivation: A pan-genome graph represents a collection of genomes and encodes sequence variations between them./ A pan genome map represents a collection of genomes , And encode the sequence changes between them .

Abstract of paper :Motivation: A pan-genome graph represents a collection of genomes and encodes sequence variations between them. It is a powerful data structure for studying multiple similar genomes. Sequence-to-graph alignment is an essential step for the construction and the analysis of pan-genome graphs. However, existing algorithms incur runtime proportional to the product of sequence length and graph size, making them inefficient for aligning long sequences against large graphs. Results: We propose the graph wavefront alignment algorithm (Gwfa), a new method for aligning a sequence to a sequence graph. Although the worst-case time complexity of Gwfa is the same as the existing algorithms, it is designed to run faster for closely matching sequences, and its runtime in practice often increases only moderately with the edit distance of the optimal alignment. On four real datasets, Gwfa is up to four orders of magnitude faster than other exact sequence-to-graph alignment algorithms. We also propose a graph pruning heuristic on top of Gwfa, which can achieve an additional ∼10-fold speedup on large graphs. Availability: Gwfa code is accessible at https://github.com/lh3/gwfa .

The pan genome map represents a collection of genomes , And encode the sequence changes between them . It is a powerful data structure , Used to study multiple similar genomes . Sequence and map alignment is an important step in the construction and analysis of Pan genome map . However , The running time of existing algorithms is proportional to the product of sequence length and graph size , It makes them inefficient in long sequence alignment of large graphs . result . We propose a graphic wavefront alignment algorithm (Gwfa), This is a new method of aligning sequences with sequence diagrams . Even though Gwfa The worst-case time complexity of the algorithm is the same as that of the existing algorithm , But it is designed to run faster against closely matched sequences , And its running time in practice often only increases moderately with the editing distance of the best alignment . On four real datasets ,Gwfa It is four orders of magnitude faster than other accurate sequence to graph arrangement algorithms . We also Gwfa A heuristic graph pruning algorithm is proposed based on , It can realize additional ∼10 Double the speed .Gwfa The code can be found in https://github.com/lh3/gwfa obtain .

The paper :CLTS-GAN: Color-Lighting-Texture-Specular Reflection Augmentation for Colonoscopy

Paper title :CLTS-GAN: Color-Lighting-Texture-Specular Reflection Augmentation for Colonoscopy

Paper time :29 Jun 2022

Field Computer vision

Address of thesis https://arxiv.org/abs/2206.14951

Code implementation https://github.com/nadeemlab/CEP

Author of the paper :Shawn Mathew, Saad Nadeem, Arie Kaufman

Brief introduction of the paper :Automated analysis of optical colonoscopy (OC) video frames (to assist endoscopists during OC) is challenging due to variations in color, lighting, texture, and specular reflections./ Because of the color 、 lighting 、 Changes in texture and specular reflection , Optical colonoscopy (OC) Automatic analysis of video frames ( stay OC During this period, assist the endoscopist ) Challenging .

Abstract of paper :Automated analysis of optical colonoscopy (OC) video frames (to assist endoscopists during OC) is challenging due to variations in color, lighting, texture, and specular reflections. Previous methods either remove some of these variations via preprocessing (making pipelines cumbersome) or add diverse training data with annotations (but expensive and time-consuming). We present CLTS-GAN, a new deep learning model that gives fine control over color, lighting, texture, and specular reflection synthesis for OC video frames. We show that adding these colonoscopy-specific augmentations to the training data can improve state-of-the-art polyp detection/segmentation methods as well as drive next generation of OC simulators for training medical students. The code and pre-trained models for CLTS-GAN are available on Computational Endoscopy Platform GitHub https://github.com/nadeemlab/CEP .

Because of the color 、 lighting 、 Changes in texture and specular reflection , Optical colonoscopy (OC) Automatic analysis of video frames ( stay OC During this period, assist the endoscope doctor ) Challenging . Previous methods either removed some of these changes through pretreatment ( Make the processing line cumbersome ), Or add different training data with annotations ( But expensive and time-consuming ). We proposed CLTS-GAN, A new deep learning model , Yes OC Color of video frame 、 lighting 、 Fine control of texture and specular synthesis . We show that , Add these colon specific enhancements to the training data , It can improve the most advanced polyp detection / Segmentation method , And promote the next generation colonoscopy simulator for training medical students .CLTS-GAN The code and pre training model are available on the computational endoscope platform GitHub https://github.com/nadeemlab/CEP Get on .

We are ShowMeAI, Dedicated to communication AI High quality content , Share industry solutions , Accelerate every technological growth with knowledge ! Click to see List of historical articles , Subscribe to topics in the official account #ShowMeAI Information daily , Can receive the latest daily push . Click on Thematic collection & Electronic monthly Quickly browse the complete works of each topic . Click on here Reply key daily Free access AI Electronic monthly and information package .

原网站

版权声明
本文为[ShowMeAI]所创,转载请带上原文链接,感谢
https://yzsam.com/2022/188/202207061646261780.html