# Autoencoders: Deep Learning with TensorFlow’s Eager Execution

Deep Learning has revolutionized the Machine Learning scene in the last years. Can we apply it to image compression? How well can a Deep Learning algorithm reconstruct pictures of kittens? What’s an autoencoder?

**By Luciano Strika, MercadoLibre**

Deep Learning has revolutionized the Machine Learning scene in the last years. Can we apply it to image compression? How well can a Deep Learning algorithm reconstruct pictures of kittens? What’s an autoencoder?

Today we’ll find the answers to all of those questions.

### Image Compression: all about the patterns

I’ve talked about Unsupervised Learning before: applying Machine Learning to discover **patterns** in **unlabelled data**.

In the case of Image Compression, it makes a lot of sense to assume most images are **not completely random**.

In more proper words, it is safe to assume most images are not completely made of **noise** (like the static when you turn on an old TV), but rather follow *some* **underlying structure**.

It would be even better if we knew all the images will have things in common. One of the most rudimentary ways of doing image compression is finding repeating streams of pixels, assigning them a (lighter) label, and replacing them with that label.

Ideally, if two images are almost the same, a compressor could store both taking barely more space than either of them. That should apply to parts of the image as well. Finally, I would guess a pattern-finding machine like Unsupervised Learning should be good at this.

For the purposes of this work, we’ll be using a dataset of pictures of kittens I downloaded from Kaggle. That will guarantee there is some common structure to most pictures.

Here’s the GitHub project with the code and data for today, in case you want to follow along. As usual however, I’ll be adding the most important snippets here.

But before we get our hands dirty with code, it’s time to learn some Deep Learning!

### Autoencoders: Unsupervised-*ish* Deep Learning

Even though we call Autoencoders “Unsupervised Learning”, they’re actually a Supervised Learning Algorithm in disguise. I know, I was shocked too!

They do however have a very peculiar property, which makes them stand out from normal classifiers: **their input and output are the same**.

### Training an Autoencoder

When we train an Autoencoder, we’ll actually be training an Artificial Neural Network that

- Takes an input vector
*X*. - Applies some math to it (I won’t get into the specifics of Deep Learning right now, but this is the book I used to learn these subjects).
- Returns another vector. In this particular case, we want it to be X again, or as close as possible.

However, there’s a trick: in the feed-forward process (going from input to output) the Neural Network will **reduce the input’s dimensionality**, and increase it again before returning the last layer.

### The curse of dimensionality

The Autoencoder takes a vector *X* as input, with potentially a lot of components. For instance, for a 3 channels –*RGB*– picture with a 48×48 resolution, *X* would have 6912 components. Even if each of them is just a float, that’s 27Kb of data for each (very small!) image. That adds up fast. The time it takes to do all the mathematical operations on so many elements is also very relevant.

We will thus have to map that vector to a **smaller space**. In this particular case, I tried reducing it to half of its original dimensions.

We intend, however, to keep most of the data.

### Architecture of an Autoencoder

We call the layer of the Network that reduces the input vector’s dimensions the *encoder*, because it *encodes* the data into a smaller vector.

We then operate on that smaller vector, in our case reducing its dimensions by a factor of 2 again. This would be the version of the input that we would store if we were designing an actual image compressor, saving 75% space. However during training, that is just an intermediate step before *decoding* it.

Finally, using that reduced vector, the Autoencoder will have to **reconstruct the original image** as well as it can.

In order to do this, it will first make it go through a *decoder* layer, which again maps the reduced vector to one with half of the input vector’s dimensions. And then a final layer, where we obtain a vector of the same dimensions as the original image, hopefully very similar.

Here’s what an Autoencoder’s layer architecture looks like:

### Results of using an Autoencoder

Training an Autoencoder will have many effects:

- The reconstructed images won’t be exactly similar to the original ones.
- We’ll be able to save the
**much lighter**reduced vector, and then use the reconstruction, provided it’s good enough (however we define that), instead of the original image. - We’ll now have a Neural Network that has learned some of the
**underlying patterns**in pictures of kittens.

This last one is the most valuable result.

Some of the industry applications for Autoencoders are:

**Feature Engineering**for Supervised Deep Learning models: instead of feeding them with data directly, we’ll feed them with the Autoencoder’s hidden layer.**Anomaly Detection**: The Autoencoder will be very bad at reconstructing pictures of dogs, landscapes or bugs. This gives us a way to check if a picture is effectively a kitten automatically.

Now that you know why we’re doing what we’re doing, let’s get our hands dirty with some actual code!

### Training an Autoencoder with Tensorflow

For this tutorial we’ll be using Tensorflow’s eager execution API. I have to say, it is a lot more intuitive than that old *Session* thing, so much so that I wouldn’t mind if there had been a drop in performance (which I didn’t perceive).

Our dataset contains over 8000 images of kittens, of various sizes and colors, and in different positions. It is available for you to look at at the GitHub project, along with the code that loads it.

Due to PyPlot’s way of handling numpy float arrays, and to accelerate convergence for the network, the images are loaded as an array of floats ranging from 0 to 1, instead of 0 to 255.

# 10% of the data will automatically be used for validation validation_size = 0.1 img_size = 48 # resize images to be 48x48 num_channels = 3 # RGB sample_size = 8192 #We'll use 8192 pictures (2**13) data = dataset.read_train_sets(train_path, img_size, ['cats'], validation_size=validation_size, sample_size=sample_size) x_train, _, _, _ = data.train.next_batch(7373) #no need to load them in batches x_valid, _, _, _ = data.valid.next_batch(819) # now that tensorflow can do it x_train[0] # array([[[0.6784314 , 0.6784314 , 0.7019608 ], ...)

### Designing the Autoencoder

Before Tensorflow swallowed Keras and became eager, writing a Neural Network with it was quite cumbersome. Now, its API has become **intuitive**.

Here’s the first Autoencoder I designed using Tensorflow’s Keras API.

from tensorflow import keras from tensorflow.keras import layers total_pixels = img_size * img_size * 3 translator_factor = 2 translator_layer_size = int(total_pixels/translator_factor) middle_factor = 2 middle_layer_size = int(translator_layer_size/middle_factor) inputs = keras.Input(shape=(img_size,img_size,3), name='cat_image') x = layers.Flatten(name = 'flattened_cat')(inputs) #turn image to vector. x = layers.Dense(translator_layer_size, activation='relu', name='encoder')(x) x = layers.Dense(middle_layer_size, activation='relu', name='middle_layer')(x) x = layers.Dense(translator_layer_size, activation='relu', name='decoder')(x) outputs = layers.Dense(total_pixels, activation='relu', name='reconstructed_cat')(x) outputs = layers.Reshape((img_size,img_size,3))(outputs) model = keras.Model(inputs=inputs, outputs=outputs)

However, this model had some major flaws! First of all, it was converging too slowly, even with a very high learning rate (we’ll get there). Secondly, it was returning a vector with the right size… except some of the elements were **higher than 1**! That didn’t mesh well with PyPlot’s graphs, nor with our inputs’ priors -each pixel has three values between 0 and 1-.

I fixed that problem by replacing the **ReLu activation function** on the last layer with a **sigmoid** activation function, so it always returned numbers between 0 and 1.

However, there was still the factor of convergence speed. What could I do to solve that? Luckily, I’ve read a few Deep Learning papers here and there, and there’s this thing I’ve used a few times called **Batch Normalization**. It deserves its own article, so I won’t be delving too deep on it.

### Batch Normalization

The short version is, a batch normalization layer calculates the mean and standard deviation for the previous layer, for the current batch of training instances. It then subtracts the mean and divides by the standard deviation, thus **normalizing the layer’s output** (for the batch). In feed-forwards after the training, the layer will start normalizing using parameters estimated from the whole training dataset.

This makes the model converge a lot faster, since it becomes less sensitive to changes in the distribution of the inputs, or the hidden layers. It will also become more **robust** to natural **distribution changes** in the inputs, being more sustainable in the long term on a productive environment.

from tensorflow import keras from tensorflow.keras import layers total_pixels = img_size * img_size * 3 translator_factor = 2 translator_layer_size = int(total_pixels/translator_factor) middle_factor = 2 middle_layer_size = int(translator_layer_size/middle_factor) inputs = keras.Input(shape=(img_size,img_size,3), name='cat_image') x = layers.Flatten(name = 'flattened_cat')(inputs) #49152 #3072 x = layers.BatchNormalization()(x) x = layers.Dense(translator_layer_size, activation='relu', name='encoder')(x) x = layers.Dense(middle_layer_size, activation='relu', name='middle_layer')(x) x = layers.BatchNormalization()(x) x = layers.Dense(translator_layer_size, activation='relu', name='decoder')(x) outputs = layers.Dense(total_pixels, activation='sigmoid', name='reconstructed_cat')(x) outputs = layers.Reshape((img_size,img_size,3))(outputs) model = keras.Model(inputs=inputs, outputs=outputs)

This way, the decoder just needs to learn how to decode centered variables with a variance of 1, and not just any random thing. The same thing applies to the encoder. This change alone made the model start converging in the first 5 or 6 epochs.

### Training the Autoencoder

It’s the moment we were all waiting for! Time to train our Autoencoder!

Tensorflow’s eager execution requires we first compile the model.

This means we define an optimizer (I’m using Adam, it’s fast), a loss (in this case, **mean squared error**, which is a pretty standard way to measure reconstruction error), and monitoring metrics. Most importantly, here is where we will choose the model’s learning rate.

customAdam = keras.optimizers.Adam(lr=0.001) #you have no idea how many times I changed this number model.compile(optimizer=customAdam, # Optimizer # Loss function to minimize loss="mean_squared_error", # List of metrics to monitor metrics=["mean_squared_error"])

And finally, we start training the model.

print('# Fit model on training data') history = model.fit(x_train, x_train, #we pass it the same input data as desired output #If the model is taking forever to train, make this bigger #If it is taking forever to load for the first epoch, make this smaller batch_size=256, epochs=10, # We pass it validation data to # monitor loss and metrics # at the end of each epoch validation_data=(x_valid, x_valid))

This will begin the training (and make my laptop sound like a Victorian furnace) and print messages like this:

Epoch 2/2 7373/7373 [==============================] - 59s 8ms/sample - loss: 0.0183 - mean_squared_error: 0.0183 - val_loss: 0.0350 - val_mean_squared_error: 0.0350

We will know the model is learning if *val_mean_squared_error*, the MSE on the validation dataset, keeps converging towards 0. If the train loss keeps dropping but the validation loss is stable or, worse, going up, then that means that the model is **overfitting**. What do we do then? We can

- add more data,
- make the compression smaller, or
- increase the learning rate.

This wasn’t a problem I had too often.

On the opposite side of the spectrum, the training loss may be bad.

How bad? Well, the value we’d get if we sampled random numbers between 0 and 1 would be around 25%, since that’s the square of 1/2, which is the average random error. That would mean we’re **underfitting**.

For most configurations I tried, that was the training loss for a long while (for instance, without using batch normalization, or when I sampled less than half of the data).

If that happens, we should decrease the learning rate (to avoid converging to a local minimum), increase the training dataset size, or be less ambitious and compress the data less.

I first tried compressing to a ninth of the original weight instead of a fourth, and there was no way the network learned that.

### The results

And finally, here are some of the results!

The model was particularly good with black or white cats, probably because they had the most contrast, and most extreme colors.

This is what happened when I tried to compress the input to a ninth of its original size:

It basically only got the outlines right, and it only worked on black or dark-grey cats.

### Conclusions

I was surprised with the results: compressing the image to a fourth of its size with the cat still being recognizable, means a classifier could probably tell there was a cat in the picture.

Tensorflow’s current API is a lot more **comfortable and intuitive** than the old one, and I’m glad I can finally do deep learning without thinking of sessions and graphs.

I think batch-normalization proved to be quite effective at accelerating the training, and it’s a tool I should use more often.

Lastly, in the future I’d like to try different image compression algorithms. K-Means Clustering based ones, and maybe a Convolutional Autoencoder. However, that’s an idea for a different article.

What about you? Think you can train a better autoencoder for this dataset? What are some of the 1000 things I should’ve done better, or hyperparameters I forgot to tweak and could’ve improved the model’s performance greatly? Let me know in the comments, I’m still very much learning.

I hope you’ve found this article entertaining or helpful, and maybe that you’ve learned something along the way.

If you wish to learn more about Deep Learning and become a professional in it, you can’t go wrong with Goodfellow and Bengio’s Deep Learning book.

Its first chapters made Deep Learning finally click for me, and the one on Autoencoders is pure gold.

*Follow me on Medium, dev.to or Twitter for more Machine Learning tutorials*.

*Thanks to Ludovic Benistant.*

**Bio: Luciano Strika** is a computer science student at Buenos Aires University, and a data scientist at MercadoLibre. He also writes about machine learning and data on www.datastuff.tech.

Original. Reposted with permission.

**Related:**

- Deep Learning, The Curse of Dimensionality, and Autoencoders
- Variational Autoencoders Explained in Detail
- Interpolation in Autoencoders via an Adversarial Regularizer