DINO: Emerging Properties in Self-Supervised Vision Transformers (Facebook AI Research Explained)
Video Statistics and Information
Channel: Yannic Kilcher
Views: 54,006
Rating: undefined out of 5
Keywords: deep learning, machine learning, arxiv, explained, neural networks, ai, artificial intelligence, paper, deep learning tutorial, what is deep learning, introduction to deep learning, facebook, facebook ai, fair, byol, swav, self supervised learning, unsupervised feature learning, unsupervised machine learning, feature engineering, stop gradient, dino, self distillation, self-distillation, segmentation maps, visual transformer, visual transformer self supervised, imagenet
Id: h3ij3F3cPIk
Channel Id: undefined
Length: 39min 12sec (2352 seconds)
Published: Sat May 01 2021
Please note that this website is currently a work in progress! Lots of interesting data and statistics to come.
Not sure this is AGI-related.
(traditionally) convolutional NNs could perform semantic scene segmentation, only after having been trained on datasets that were pre-segmented. In most cases, those training sets had to be manually segmented by actual humans drawing on images. That is cumbersome and expensive.
With DINO, object segmentation is automated, allowing the creation of terabytes of pre-segmented datasets for further training downstream. That's great news for computer vision and machine learning. However, I don't see any obvious connection to AGI.