Skip to content

Latest commit

 

History

History
159 lines (138 loc) · 8.57 KB

index.md

File metadata and controls

159 lines (138 loc) · 8.57 KB
title layout header excerpt intro feature_row organizers_row speakers_row
<span class='header-marking'>2nd Visual Inductive Priors for Data-Efficient Deep Learning Workshop</span>
splash
overlay_image
assets/images/header_background_centered.png
<span class='header-marking'>ICCV 2021</span><br/><span class='header-marking'>11 October 2021, afternoon</span>
excerpt
Saving data by adding visual knowledge priors to Deep Learning.
image_path alt title excerpt url btn_label btn_class
assets/images/callforpapers.png
placeholder image 2
Call for papers
We showcase recent work on data-efficient computer vision: [OpenReview site](https://openreview.net/group?id=thecvf.com/ICCV/2021/Workshop/VIPriors).
call-for-papers
Call for Papers
btn--primary
image_path alt title excerpt url btn_label btn_class
assets/images/poster.png
placeholder image 2
Present a poster
We invite researchers to present their recent published works on data-efficient computer vision as a poster at our workshop. This may include works published at the main ICCV 2021 conference paper track.
call-for-papers/#call-for-posters
Call for Posters
btn--primary
image_path title excerpt url btn_label btn_class
assets/images/challenge.png
VIPriors challenges
We host five data efficieny challenges on action recognition, classification, detection, segmentation and object tracking.
challenges
Final rankings
btn--primary
image_path alt title excerpt url btn_label btn_class
assets/images/JanVanGemert.jpg
Jan van Gemert
Jan van Gemert
Delft University of Technology
Website
btn--default
image_path alt title excerpt url btn_label btn_class
assets/images/MatthiasBethge.jpg
Matthias Bethge
Matthias Bethge
Bethge Lab
Website
btn--default
image_path alt title excerpt
assets/images/NergisTomen.jpg
Nergis Tömen
Nergis Tömen
Delft University of Technology
image_path alt title excerpt url btn_label btn_class
assets/images/AttilaLengyel.jpg
Attila Lengyel
Attila Lengyel
Delft University of Technology
Website
btn--default
image_path alt title excerpt url btn_label btn_class
assets/images/Robert-JanBruintjes.jpg
Robert-Jan Bruintjes
Robert-Jan Bruintjes
Delft University of Technology
Website
btn--default
image_path alt title excerpt
assets/images/OsmanKayhan.png
Osman Semih Kayhan
Osman Semih Kayhan
Bosch Security Systems B.V.
image_path alt title excerpt
assets/images/MarcosBaptistaRios.jpg
Marcos Baptista Ríos
Marcos Baptista Ríos
Gradiant
image_path alt title excerpt url btn_label btn_class
/assets/images/AnimaAnandkumar.png
Anima Anandkumar
Anima Anandkumar
Caltech, NVIDIA
Website
btn--default
image_path alt title excerpt url btn_label btn_class
/assets/images/EkinCubuk.jpg
Ekin Dogus Cubuk
Ekin Dogus Cubuk
Google Research
Website
btn--default
image_path alt title excerpt url btn_label btn_class
/assets/images/ChelseaFinn.jpg
Chelsea Finn
Chelsea Finn
Stanford University
Website
btn--default
image_path alt title excerpt url btn_label btn_class
/assets/images/MaxWelling.jpg
Max Welling
Max Welling
University of Amsterdam, Qualcomm
Website
btn--default

{% include feature_row id="intro" type="center" %}

{% include feature_row %}

Watch the recording

{% include video id="KBNjPtcVKz0" provider="youtube" %}

{% include twitter_sidebar %}

Thanks to all for attending

We enjoyed a very exciting workshop at ICCV 2021. We thank all presenters for their efforts and all participants for their attention. This website will keep a record of all presented materials. We hope to see you all next year (venue TBD) for the next workshop!

About the workshop

Data is fueling deep learning, yet it is costly to gather and to annotate. Training on massive datasets has a huge energy consumption adding to our carbon footprint. In addition, there are only a select few deep learning behemoths which have billions of data points and thousands of expensive deep learning hardware GPUs at their disposal. This workshop focuses on how to pre-wire deep networks with generic visual inductive innate knowledge structures, which allows to incorporate hard won existing generic knowledge. Visual inductive priors are data efficient: what is built-in no longer has to be learned, saving valuable training data.

Excellent recent research investigates data efficiency in deep networks by exploiting other data sources through unsupervised learning, re-using existing datasets, or synthesizing artificial training data. However, not enough attention is given on how to overcome the data dependency by adding prior knowledge to deep nets. As a consequence, all knowledge has to be (re-)learned implicitly from data, making deep networks hard to understand black boxes which are susceptible to dataset bias requiring huge datasets and compute resources. This workshop aims to remedy this gap by investigating how to flexibly pre-wire deep networks with generic visual innate knowledge structures, which allows to incorporate hard won existing knowledge from physics such as light reflection or geometry.

The great power of deep neural networks is their incredible flexibility to learn. The direct consequence of such power, is that small datasets can simply be memorized and the network will likely not generalize to unseen data. Regularization aims to prevent such over-fitting by adding constraints to the learning process. Much work is done on regularization of internal network properties and architectures. In this workshop we focus on regularization methods based on innate priors. There is strong evidence that an innate prior benefits deep nets: adding convolution to deep networks yields a convolutional deep neural network (CNN) which is hugely successful and has permeated the entire field. While convolution was initially applied on images, it is now generalized to graph networks, speech, language, 3D data, video, etc. Convolution models translation invariance in images: an object may occur anywhere in the image, and thus instead of learning parameters at each location in the image, convolution allows to only consider local relations, yet, share parameters over all image locations. This allows a strong reduction in both number of parameters and examples to learn from. This workshop aims to further the great success of convolution, exploiting innate regularizing structures yielding a significant reduction of training data.

This workshop is organized in collaboration with SynergySports. SynergySports is co-organizing the VIPriors 2021 challenges. Head over to the challenges page to find out more!

Workshop program

| EDT (UTC-4) | UTC | CEST (UTC+2) | CST (UTC+8) | | | | -- | -- | -- | -- | -- | | 13:00 | 17:00 | 19:00 | 01:00 | Opening | Challenge winners will be announced. | | 13:15 | 17:15 | 19:15 | 01:15 | Invited talk: Chelsea Finn | | | 13:45 | 17:45 | 19:45 | 01:45 | Invited talk: Max Welling | | | 14:15 | 18:15 | 20:15 | 02:15 | Invited talks Q&A | | | 14:30 | 18:30 | 20:30 | 02:30 | Break | | | 14:40 | 18:40 | 20:40 | 02:40 | Oral presentations | Three 10-minute presentations. | | 15:10 | 19:10 | 21:10 | 03:10 | Poster session | Posters, live discussion on Gatherly. | | 16:00 | 20:00 | 22:00 | 04:00 | Invited talk: Anima Anandkumar | Slides are available here. | | 16:30 | 20:30 | 22:30 | 04:30 | Invited talk: Ekin Dogus Cubuk | | | 17:00 | 21:00 | 23:00 | 05:00 | Invited talks Q&A | | | 17:15 | 21:15 | 23:15 | 05:15 (est.) | Conclusion | |

Invited speakers

{% include feature_row id="speakers_row" %}

Organizers

{% include feature_row id="organizers_row" %}

Contact

Email us at vipriors-ewi AT tudelft DOT nl