Silver BlogTop 10 Computer Vision Papers 2020

The top 10 computer vision papers in 2020 with video demos, articles, code, and paper reference.



By Louis (What's AI) Bouchard, Montrealer, explaining AI stuff on YouTube and Medium

Image for post

Even with everything that happened in the world this year, we still had the chance to see a lot of amazing research come out. Especially in the field of artificial intelligence and more precisely computer vision. More, many important aspects were highlighted this year, like the ethical aspects, important biases, and much more. Artificial intelligence and our understanding of the human brain and its link to AI is constantly evolving, showing promising applications in the soon future, which I will definitely cover.

Here are my top 10 of the most interesting research papers of the year in computer vision, in case you missed any of them. In short, it is basically a curated list of the latest breakthroughs in AI and CV with a clear video explanationlink to a more in-depth article, and code (if applicable). Enjoy the read, and let me know if I missed any important papers in the comments, or by contacting me directly on LinkedIn!

The complete reference to each paper is listed at the end of this article.

Access the complete list in a GitHub repository

Tag me on Twitter (@Whats_AI) or LinkedIn (Louis (What’s AI) Bouchard) if you share the article!

 

Watch a complete computer vision 2020 rewind in 5 minutes

 

 

Sea-thru: A Method For Removing Water From Underwater Images [1]

 
Have you ever wondered how the ocean would look like without water Remove this blue-green tint of the underwater pictures, and still have the true colors of a coral reef? Well, using computer vision and machine learning algorithms, researchers from the University of Haifa were able to accomplish exactly that!

This AI removes the water from underwater images!
Have you ever wondered how the ocean would look like without water? Researchers recently achieved that by using…
 

Click here for the Sea-thru code

 

Neural circuit policies enabling auditable autonomy [2]

 

Researchers from IST Austria and MIT have successfully trained a self-driving car using a new artificial intelligence system based on the brains of tiny animals, such as threadworms. They achieved that with only a few neurons able to control the self-driving car, compared to the millions of neurons needed by the popular deep neural networks such as Inceptions, Resnets, or VGG. Their network was able to completely control a car using only 75 000 parameters, composed of 19 control neurons, rather than millions!

A New Brain-inspired Intelligent System Drives a Car Using Only 19 Control Neurons!
Imitating the nematode’s nervous system to process information efficiently, this new intelligent system is more robust…
 

NeRV: Neural Reflectance and Visibility Fields
for Relighting and View Synthesis
 [3]

 
This new method is able to generate a complete 3-dimensional scene and has the ability to decide the lighting of the scene. All this with very limited computation costs and amazing results compared to previous approaches.

Generate a Complete 3D Scene Under Arbitrary Lighting Conditions from a Set of Input Images
This new method is able to generate a complete 3-dimensional scene and has the ability to decide the lighting of the…
 

YOLOv4: Optimal Speed and Accuracy of Object Detection [4]

 
This 4th version has been recently introduced in April 2020 by Alexey Bochkovsky et al. in the paper “YOLOv4: Optimal Speed and Accuracy of Object Detection”. The main goal of this algorithm was to make a super-fast object detector with high quality in terms of accuracy.

The YOLOv4 algorithm | Introduction to You Only Look Once, Version 4 | Real-Time Object Detection
I recently made a post explaining the basics of the initial You Only Look Once, also known as the YOLO algorithm. And…
 

PULSE: Self-Supervised Photo Upsampling via Latent Space Exploration of Generative Models [5]

 
This new algorithm transforms a blurry image into a high-resolution image!

It can take a super low-resolution 16x16 image and turn it into a 1080p high definition human face! You don’t believe me? Then you can do just like me and try it on yourself in less than a minute! But first, let’s see how they did that.

This AI makes blurry faces look 60 times sharper
This new algorithm transforms a blurry image into a high-resolution image! It can take a super low-resolution 16x16 image…
 

Image GPT — Generative Pretraining from Pixels [6]

 
A good AI, like the one used in Gmail, can generate coherent text and finish your phrase. This one uses the same principles in order to complete an image! All done in an unsupervised training with no labels required at all!

This AI Can Generate the Other Half of a Picture Using a GPT Model
A good AI, like the one used in Gmail, can generate coherent text and finish your phrase. This one uses the same…
 

DeepFaceDrawing: Deep Generation of Face Images from Sketches [7]

 
You can now generate high-quality face images from rough or even incomplete sketches with zero drawing skills using this new image-to-image translation technique! If your drawing skills as bad as mine you can even adjust how much the eyes, mouth, and nose will affect the final image! Let’s see if it really works and how they did it.

AI Generates Real Faces From Sketches!
You can now generate high-quality face images from rough or even incomplete sketches with zero drawing skills using…
 

PIFuHD: Multi-Level Pixel-Aligned Implicit Function for High-Resolution 3D Human Digitization [8]

 
This AI Generates 3D high-resolution reconstructions of people from 2D images! It only needs a single image of you to generate a 3D avatar that looks just like you, even from the back!

 
AI Generates 3D high-resolution reconstructions of people from 2D images | Introduction to PIFuHD
This AI Generates 3D high-resolution reconstructions of people from 2D images! It only needs a single image of you to…
 

RAFT: Recurrent All-Pairs Field Transforms for Optical Flow [9]

 
ECCV 2020 Best Paper Award Goes to Princeton Team. They developed a new end-to-end trainable model for optical flow. Their method beats state-of-the-art architectures’ accuracy across multiple datasets and is way more efficient. They even made the code available for everyone on their Github!

ECCV 2020 Best Paper Award | A New Architecture For Optical Flow
ECCV 2020 Best Paper Award Goes to Princeton Team. They developed a new end-to-end trainable model for optical flow…
 

Click here for the RAFT code

 

Learning Joint Spatial-Temporal Transformations for Video Inpainting [10]

 
This AI can fill the missing pixels behind a removed moving object and reconstruct the whole video with way more accuracy and less blurriness than current state-of-the-art approaches!

This AI takes a video and fills the missing pixels behind an object!
Video Inpainting — Microsoft Research
 

Click here for this Video Inpainting code

 

Old Photo Restoration via Deep Latent Space Translation [Bonus 1]

 
Imagine having the old, folded, and even torn pictures of your grandmother when she was 18 years old in high definition with zero artifacts. This is called old photo restoration and this paper just opened a whole new avenue to address this problem using a deep learning approach.

Old Photo Restoration using Deep Learning
Imagine having the old, folded, and even torn pictures of your grandmother when she was 18 years old in high definition…
 

Click here for the Old Photo Restoration code

 

Is a Green Screen Really Necessary for Real-Time Portrait Matting? [Bonus 2]

 
Human matting is an extremely interesting task where the goal is to find any human in a picture and remove the background from it. It is really hard to achieve due to the complexity of the task, having to find the person or people with the perfect contour. In this post, I review the best techniques used over the years and a novel approach published on November 29th, 2020. Many techniques are using basic computer vision algorithms to achieve this task, such as the GrabCut algorithm, which is extremely fast, but not very precise.

High-Quality Background Removal Without Green Screens
This new background removal technique can extract a person from a single input image, without the need for a green…
 

Click here for the MODNet code

 

DeOldify [Bonus 3]

 
DeOldify is a technique to colorize and restore old black and white images or even film footage. It was developed and is still getting updated by only one person Jason Antic. It is now the state of the art way to colorize black and white images, and everything is open-sourced, but we will get back to this in a bit.

This AI can Colorize your Black & White Photos with Full Photorealistic Renders! (DeOldify)
This method is called DeOldify and works on pretty much any picture. If you don’t believe me, you can even try it…
 

Click here for the DeOldify code

 

Conclusion

 
As you can see, this was an extremely insightful year for computer vision. I will be sure to cover the most exciting and interesting papers of 2021, and I would love it if you could take part in this adventure! If you like my work and want to stay up-to-date with AI technologies, you should definitely follow me on my social media channels.

Access the complete list in a GitHub repository

Tag me on Twitter (@Whats_AI) or LinkedIn (Louis (What’s AI) Bouchard) if you share the article!

 

If you are interested in AI research, here is another great article for you:

 
2020: A Year Full of Amazing AI Papers — A Review
A curated list of the latest breakthroughs in AI by release date with a clear video explanation, link to a more…
 

Paper references
 
[1] Akkaynak, Derya & Treibitz, Tali. (2019). Sea-Thru: A Method for Removing Water From Underwater Images. 1682–1691. 10.1109/CVPR.2019.00178.
[2] Lechner, M., Hasani, R., Amini, A. et al. Neural circuit policies enabling auditable autonomy. Nat Mach Intell 2, 642–652 (2020). https://doi.org/10.1038/s42256-020-00237-3
[3] P. P. Srinivasan, B. Deng, X. Zhang, M. Tancik, B. Mildenhall, and J. T. Barron, “Nerv: Neural reflectance and visibility fields for relighting and view synthesis,” in arXiv, 2020.
[4] A. Bochkovskiy, C.-Y. Wang, and H.-Y. M. Liao, Yolov4: Optimal speed and accuracy of object detection, 2020. arXiv:2004.10934 [cs.CV].
[5] S. Menon, A. Damian, S. Hu, N. Ravi, and C. Rudin, Pulse: Self-supervised photo upsampling via latent space exploration of generative models, 2020. arXiv:2003.03808 [cs.CV].
[6] M. Chen, A. Radford, R. Child, J. Wu, H. Jun, D. Luan, and I. Sutskever, “Generative pretraining from pixels,” in Proceedings of the 37th International Conference on Machine Learning, H. D. III and A. Singh, Eds., ser. Proceedings of Machine Learning Research, vol. 119, Virtual: PMLR, 13–18 Jul 2020, pp. 1691–1703. [Online].
[7] S.-Y. Chen, W. Su, L. Gao, S. Xia, and H. Fu, “DeepFaceDrawing: Deep generation of face images from sketches,” ACM Transactions on Graphics (Proceedings of ACM SIGGRAPH2020), vol. 39, no. 4, 72:1–72:16, 2020. Available:http://proceedings.mlr.press/v119/chen20s.html.
[8] S. Saito, T. Simon, J. Saragih, and H. Joo, Pifuhd: Multi-level pixel-aligned implicit function for high-resolution 3d human digitization, 2020. arXiv:2004.00452 [cs.CV].
[9] Z. Teed and J. Deng, Raft: Recurrent all-pairs field transforms for optical flow, 2020. arXiv:2003.12039 [cs.CV].
[10] Y. Zeng, J. Fu, and H. Chao, Learning joint spatial-temporal transformations for video in-painting, 2020. arXiv:2007.10247 [cs.CV].
[Bonus 1] Z. Wan, B. Zhang, D. Chen, P. Zhang, D. Chen, J. Liao, and F. Wen, Old photo restoration via deep latent space translation, 2020. arXiv:2009.07047 [cs.CV].
[Bonus 2] Z. Ke, K. Li, Y. Zhou, Q. Wu, X. Mao, Q. Yan, and R. W. Lau, “Is a green screen really necessary for real-time portrait matting?” ArXiv, vol. abs/2011.11961, 2020.
[Bonus 3] Jason Antic, Creator of DeOldify, https://github.com/jantic/DeOldify

 
Original. Reposted with permission.

Related: