If you were on YouTube trying to learn about variational autoencoders (VAEs) as I was, you might have come across Ahlad Kumar’s series on the topic. In his second video (embedded above), he explained KL divergence which we will later see is in fact a building block of the loss function in the VAE. This article aims to bridge ideas in probability theory as you may have learnt in school to those in the video. We will then re-look at the proof for KL divergence between 2 multivariate Gaussians (a.k.a normal distributions).
In this article, I will be sharing my takeaways from reading Background Matting: The World is Your Green Screen by Sengupta et.al.
Visual effects in movies make use of a green screen to superimpose computer-generated landscapes into the background. That’s green-screen matting.
Writing this article after reading Fixing the train-test resolution discrepancy (Touvron et. al). I will include the relevant sections of the paper in brackets. I have also used images from the paper.
The current best practice for image preprocessing to train image classifiers looks a little something like this: (Section 1)
train_transform = transforms.Compose([
])test_transform = transforms.Compose([
We take different image preprocessings steps at train and test time. And here’s why. …
Greetings! In this article, I will be consolidating my thoughts on Jeremy Howard’s implementation of Kaiming Init from fastai’s Lesson 9. When I encountered this implementation, I had a few troubles in my understanding:
The code for Kaiming Init in the fully connected layer (as presented in Lesson 8) is very different from that in the convolutional layer. I guess this was what threw me off a bit.