Overview and benchmark of traditional and deep learning models in text classification
In this post, traditional and deep learning models in text classification will be thoroughly investigated, including a discussion into both Recurrent and Convolutional neural networks.
By Ahmed Besbes, Capgemini Consulting.
This article is an extension of a previous one I wrote when I was experimenting sentiment analysis on twitter data. Back in the time, I explored a simple model: a twolayer feedforward neural network trained on keras. The input tweets were represented as document vectors resulting from a weighted average of the embeddings of the words composing the tweet.
The embedding I used was a word2vec model I trained from scratch on the corpus using gensim. The task was a binary classification and I was able with this setting to achieve 79% accuracy.
The goal of this post is to explore other NLP models trained on the same dataset and then benchmark their respective performance on a given test set.
We'll go through different models: from simple ones relying on a bagofword representation to a heavy machinery deploying convolutional/recurrent networks: We'll see if we'll score more than 79% accuracy!
I will start from simple models and add up complexity progressively. The goal is also to show that simple models work well too.
So I'm going to try out these:
 Logistic regression with word ngrams
 Logistic regression with character ngrams
 Logistic regression with word and character ngrams
 Recurrent neural network (bidirectional GRU) without pretrained embeddings
 Recurrent neural network (bidirectional GRU) with GloVe pretrained embeddings
 Multi channel Convolutional Neural Network
 RNN (Bidirectional GRU) + CNN model
By the end of this post, you will have a boilerplate code for each of these NLP techniques. It'll help you kickstart your NLP project and eventually achieve stateofthe art results (some of these models are really powerful).
We'll also provide a comprehensive benchmark from which we'll tell which model is best suited for predicting the sentiment of a tweet.
In the related git repo, I will publish the different models, their predictions, as well as the test set. You can try them yourself and get confident about the results
Let's get started!
import os import re import warnings warnings.simplefilter("ignore", UserWarning) from matplotlib import pyplot as plt %matplotlib inline import pandas as pd pd.options.mode.chained_assignment = None import numpy as np from string import punctuation from nltk.tokenize import word_tokenize from sklearn.model_selection import train_test_split from sklearn.feature_extraction.text import TfidfVectorizer from sklearn.linear_model import LogisticRegression from sklearn.metrics import accuracy_score, auc, roc_auc_score from sklearn.externals import joblib import scipy from scipy.sparse import hstack
0  Data preprocessing
The dataset can be downloaded from this link.
We'll load it and restrict ourselves to the variables we need (Sentiment and SentimentText).
It contains 1578614 classified tweets, each row is marked as 1 for positive sentiment and 0 for negative sentiment.
The author recommend using 1/10 for testing the algorithm and the rest for training.
data = pd.read_csv('./data/tweets.csv', encoding='latin1', usecols=['Sentiment', 'SentimentText']) data.columns = ['sentiment', 'text'] data = data.sample(frac=1, random_state=42) print(data.shape)
for row in data.head(10).iterrows():
print(row[1]['sentiment'], row[1]['text'])
Tweets are noisy, let's clean them by removing urls, hashtags and user mentions.
def tokenize(tweet):
tweet = re.sub(r'http\S+', '', tweet)
tweet = re.sub(r"#(\w+)", '', tweet)
tweet = re.sub(r"@(\w+)", '', tweet)
tweet = re.sub(r'[^\w\s]', '', tweet)
tweet = tweet.strip().lower()
tokens = word_tokenize(tweet)
return tokens
Once the data is clean, we save it on disk.
data['tokens'] = data.text.progress_map(tokenize)
data['cleaned_text'] = data['tokens'].map(lambda tokens: ' '.join(tokens))
data[['sentiment', 'cleaned_text']].to_csv('./data/cleaned_text.csv')
data = pd.read_csv('./data/cleaned_text.csv')
print(data.shape)
data.head()
Now that the dataset is cleaned, let's prepare a train/test split to build our models.
We'll use this split throughout all the notebook.
x_train, x_test, y_train, y_test = train_test_split(data['cleaned_text'],
data['sentiment'],
test_size=0.1,
random_state=42,
stratify=data['sentiment'])
print(x_train.shape, x_test.shape, y_train.shape, y_test.shape)
I'm saving the test labels on disk for later use.
pd.DataFrame(y_test).to_csv('./predictions/y_true.csv',
index=False, encoding='utf8')
Let's start applying some machine learning now:
1  Bag of word model based on word ngrams
So. What's an ngram ?
As we see it in this figure, ngrams are simply all combinations of adjacent words (in this case) of length n that you can find in your source text.
In our model, we are going to use unigrams (n=1) and bigrams (n=2) as features.
The dataset will therefore be represented as a matrix where each row refers to a tweet and each column refers to a feature (unigram or bigram) extracted from the text (after tokenization and cleaning). Each cell will be the tfidf score. (we could also use a simple count but tfidf is more used in general and usually works better). We call this matrix the documentterm matrix.
As you can imagine, the number of unique unigrams and bigrams of 1.5 milliontweet corpus is huge. In practice, we will set this number to a fixed value, for a computational reason. You can determine this value by using a crossvalidation.
This is what the corpus should look like after vectorization.
I like pizza a lot
Let's say we want to feed this sentence to a predictive model using the features described above.
Given that we're using unigrams and bigrams, the model will extract the following features:
i, like, pizza, a, lot, i like, like pizza, pizza a, a lot
Therefore, the sentence will be formed by a vector of size N (= total number of tokens) containing lots of zeros and the tfidf scores of these ngrams. So you can clearly see that we will be dealing with large and sparse vectors.
When dealing with large and sparse data, linear models generally perform quite well. Besides, they are faster to train than other types of models (e.g. treebased models).
I can tell from past experience that logistic regression works well on top of sparse tfidf matrices.
vectorizer_word = TfidfVectorizer(max_features=40000,
min_df=5,
max_df=0.5,
analyzer='word',
stop_words='english',
ngram_range=(1, 2))
vectorizer_word.fit(x_train, leave=False)
tfidf_matrix_word_train = vectorizer_word.transform(x_train)
tfidf_matrix_word_test = vectorizer_word.transform(x_test)
After generating tfidf matrices for both train and test sets, we can build our first model an test it.
The tifidf matrices are the features of the logistic regression.
lr_word = LogisticRegression(solver='sag', verbose=2)
lr_word.fit(tfidf_matrix_word_train, y_train)
Once the model is trained, we apply it on the test data to get predictions. Then we save these values as well as the model on disk.
joblib.dump(lr_word, './models/lr_word_ngram.pkl')
y_pred_word = lr_word.predict(tfidf_matrix_word_test)
pd.DataFrame(y_pred_word, columns=['y_pred']).to_csv
('./predictions/lr_word_ngram.csv', index=False)
Let's see what accuracy score we have:
y_pred_word = pd.read_csv('./predictions/lr_word_ngram.csv')
print(accuracy_score(y_test, y_pred_word))
78.2% accuracy for a first model ! Quite not bad. Let's move to the next model.
2  Bag of word model based on character ngrams
We never said that ngrams were for words only. We can apply them at a character level as well.
You see it coming, right? We're going to apply the same code above to character ngrams instead, and we will go up to 4grams.
This basically means that a sentence like "I like this movie" will have these features:
I, l, i, k, e, ..., I li, lik, like, ..., this, ... , is m, s mo, movi, ...
Character ngrams are surprisingly very effective. They can even outperform word tokens in modeling a language task. For example spam filters or Native Language Identification heavily rely on character ngrams.
Unlike the previous model which learns combinations of words, this model learns combinations of letters, which can handle the morphological makeup of a word.
One of the advantages of the characterbased representation is the better handling of misspelled words.
Let's run the same pipeline :
vectorizer_char = TfidfVectorizer(max_features=40000,
min_df=5,
max_df=0.5,
analyzer='char',
ngram_range=(1, 4))
vectorizer_char.fit(tqdm_notebook(x_train, leave=False));
tfidf_matrix_char_train = vectorizer_char.transform(x_train)
tfidf_matrix_char_test = vectorizer_char.transform(x_test)
lr_char = LogisticRegression(solver='sag', verbose=2)
lr_char.fit(tfidf_matrix_char_train, y_train)
y_pred_char = lr_char.predict(tfidf_matrix_char_test)
joblib.dump(lr_char, './models/lr_char_ngram.pkl')
pd.DataFrame(y_pred_char, columns=['y_pred']).
to_csv('./predictions/lr_char_ngram.csv', index=False)
y_pred_char = pd.read_csv('./predictions/lr_char_ngram.csv')
print(accuracy_score(y_test, y_pred_char))
80.4% accuracy ! Characterngrams perfom better that wordngrams.
3  Bag of word model based on word and character ngrams
Character ngram features seem to provide a better accuracy than word ngrams. But what about the combination of the two: word + character ngrams?
Let's concatenate the two tfidf matrices we generated and build a new, hybrid tfidf matrix.
This model will help us learn the indentity of a word and its possible neighbors as well as its morphological structure.
These properties are combined.
tfidf_matrix_word_char_train =
hstack((tfidf_matrix_word_train, tfidf_matrix_char_train))
tfidf_matrix_word_char_test =
hstack((tfidf_matrix_word_test, tfidf_matrix_char_test))
lr_word_char = LogisticRegression(solver='sag', verbose=2)
lr_word_char.fit(tfidf_matrix_word_char_train, y_train)
y_pred_word_char = lr_word_char.predict(tfidf_matrix_word_char_test)
joblib.dump(lr_word_char, './models/lr_word_char_ngram.pkl')
pd.DataFrame(y_pred_word_char, columns=['y_pred']).
to_csv('./predictions/lr_word_char_ngram.csv', index=False)
y_pred_word_char = pd.read_csv('./predictions/lr_word_char_ngram.csv')
print(accuracy_score(y_test, y_pred_word_char))
Awesome: 81.4% accuracy. We just increased by one whole unit and outperformed the two previous settings.
What can we say about bagofword models before we move on?
 Pros: They can be surprisingly powerful given their simplicity, they are fast to train, and easy to understand.
 Cons: Even though ngrams bring some context between words, bag of word models fail in modeling longterm dependencies between words in a sequence.
Now we're going to dive into deep learning models. The reason deep learning outperform bag of word models is the ability to capture the sequencial dependency between words in a sentence. This has been possible thanks to the invention of special neural network architectures called Recurrent Neural Networks.
I won't cover the theoretical foundations of RNNs, but here's a link I find worth reading. It's from Cristopher Olah's blog. It details LSTM: Long Short Term Memory. A special kind of RNN.
Before starting, we have to setup a deep learning dedicated environment that uses Keras on top of Tensorflow. I honestly tried to run everything on my personal laptop but given the important size of the dataset and the complexity of RNN architectures, this has not been practical. At all.
One good option is AWS. I generally use this deep learning AMI on an EC2 p2.xlarge instance. Amazon AMI are preconfigured VM images, where all the packages (Tensorflow, PyTocrh, Keras, etc. ) are installed. I highly recommend this one which I have been using for a while.
from keras.preprocessing.text import Tokenizer
from keras.preprocessing.text import text_to_word_sequence
from keras.preprocessing.sequence import pad_sequences
from keras.models import Model
from keras.models import Sequential
from keras.layers import Input, Dense, Embedding, Conv1D,
Conv2D, MaxPooling1D, MaxPool2D
from keras.layers import Reshape, Flatten, Dropout, Concatenate
from keras.layers import SpatialDropout1D, concatenate
from keras.layers import GRU, Bidirectional,
GlobalAveragePooling1D, GlobalMaxPooling1D
from keras.callbacks import Callback
from keras.optimizers import Adam
from keras.callbacks import ModelCheckpoint, EarlyStopping
from keras.models import load_model
from keras.utils.vis_utils import plot_model
4  Recurrent Neural Network without pretrained embedding
RNNs may look scary. Although they're complex to understand, they're quite interesting. They encapsulate a very beautiful design that overcomes traditional neural networks' shortcomings that rise when dealing with sequence data: text, time series, videos, DNA sequences, etc.
An RNN is a sequence of neural network blocks that are linked to each others like a chain. Each one is passing a message to a successor.
Again if you want to dive into the internal mechanics, I highly recommend Colah's blog out of which the diagram below is taken.
We will process text data, which is a sequence type. The order of words is very important to the meaning. Hopefully RNNs take care of this and can capture longterm dependencies.
To use Keras on text data, we firt have to preprocess it. For this, we can use Keras' Tokenizer class. This object takes as argument num_words which is the maximum number of words kept after tokenization based on their word frequency.
MAX_NB_WORDS = 80000
tokenizer = Tokenizer(num_words=MAX_NB_WORDS)
tokenizer.fit_on_texts(data['cleaned_text'])
Once the tokenizer is fitted on the data, we can use it to convert text strings to sequences of numbers.
These numbers represent the position of each word in the dictionary (think of it as mapping).
Let's see an example:
x_train[15]
Here's how the tokenizer turns it into a sequence of digits.
tokenizer.texts_to_sequences([x_train[15]])
Let's now apply this tokenizer on the train and test sequences:
train_sequences = tokenizer.texts_to_sequences(x_train)
test_sequences = tokenizer.texts_to_sequences(x_test)
Now the tweets are mapped to lists of integers. However, we still cannot stack them together in a matrix since they have different lengths. Hopefully Keras allows to pad sequences with 0s to a maximum length. We'll set this length to 35. (which is the maximum numbers of tokens in the tweets).
MAX_LENGTH = 35
padded_train_sequences = pad_sequences(train_sequences, maxlen=MAX_LENGTH)
padded_test_sequences = pad_sequences(test_sequences, maxlen=MAX_LENGTH)
padded_train_sequences
padded_train_sequences.shape
Now the data is ready to be fed to an RNN.
Here are some elements of the architecture I'll be using:
 An embedding dimension of 300. This means that each word from the 80000 that we'll be using is mapped to a 300dimension dense vector (of float numbers). The mapping will adjust throughout the training.
 A spatial dropout is applied on the embedding layer to reduce overfitting: it basically looks at batches of 35x300 matrices and randomly drop (set to 0) word vectors (i.e rows) in each matrix. This helps not to focus on specific words in an attempt to generalize well.
 A bidirectional Gated Recurrent Unit (GRU): this is the recurrent network part. It's a faster variant of the LSTM architecture. Think of it as a combination of two recurrent networks that scan the text sequence in both directions: from left to right and from right to left. This allows the network, when reading a given word, to understand it by using the context from both past and future information. The GRU takes as parameter a number of units which is the dimension of the output h_t of each network block. We will set this number to 100. And since we are using a bidirectional version of the GRU, the final output per RNN block will be of dimension 200.
The output of the bidirectional GRU has the dimension (batch_size, timesteps, units). This means that if we use a typical batch size of 256, this dimension will be (256, 35, 200)
 On top of every batch, we apply a global average pooling that consists in averaging the output vectors corresponding to the each time step (i.e the words)
 We apply the same operation with max pooling.
 We concatenate the outputs of the two previous operations.
def get_simple_rnn_model():
embedding_dim = 300
embedding_matrix = np.random.random((MAX_NB_WORDS, embedding_dim))
inp = Input(shape=(MAX_LENGTH, ))
x = Embedding(input_dim=MAX_NB_WORDS,
output_dim=embedding_dim, input_length=MAX_LENGTH,
weights=[embedding_matrix], trainable=True)(inp)
x = SpatialDropout1D(0.3)(x)
x = Bidirectional(GRU(100, return_sequences=True))(x)
avg_pool = GlobalAveragePooling1D()(x)
max_pool = GlobalMaxPooling1D()(x)
conc = concatenate([avg_pool, max_pool])
outp = Dense(1, activation="sigmoid")(conc)
model = Model(inputs=inp, outputs=outp)
model.compile(loss='binary_crossentropy',
optimizer='adam',
metrics=['accuracy'])
return model
rnn_simple_model = get_simple_rnn_model()
Let's look at the different layers of this model:
plot_model(rnn_simple_model,
to_file='./images/article_5/rnn_simple_model.png',
show_shapes=True,
show_layer_names=True)
During the training, model checkpoint is used. It allows to automatically save (on disk) the best models (w.r.t accuracy measure) at the end of each epoch.
filepath="./models/rnn_no_embeddings/weights
improvement{epoch:02d}{val_acc:.4f}.hdf5"
checkpoint = ModelCheckpoint(filepath, monitor='val_acc',
verbose=1, save_best_only=True, mode='max')
batch_size = 256
epochs = 2
history = rnn_simple_model.fit(x=padded_train_sequences,
y=y_train,
validation_data=(padded_test_sequences, y_test),
batch_size=batch_size,
callbacks=[checkpoint],
epochs=epochs,
verbose=1)
best_rnn_simple_model = load_model('./models/rnn_no_embeddings/
weightsimprovement010.8262.hdf5')
y_pred_rnn_simple = best_rnn_simple_model.predict(padded_test_sequences,
verbose=1, batch_size=2048)
y_pred_rnn_simple = pd.DataFrame(y_pred_rnn_simple, columns=['prediction'])
y_pred_rnn_simple['prediction'] = y_pred_rnn_simple['prediction'].
map(lambda p: 1 if p >= 0.5 else 0)
y_pred_rnn_simple.to_csv('./predictions/y_pred_rnn_simple.csv', index=False)
y_pred_rnn_simple = pd.read_csv('./predictions/y_pred_rnn_simple.csv')
print(accuracy_score(y_test, y_pred_rnn_simple))
82.6% accuracy ! Quite not bad ! We are now performing better than the previous bagofword models because we are taking into account the sequence nature of the text.
Can we do better?
5  Recurrent Neural Network with GloVe pretrained embeddings
In the last model, the embedding matrix was initialized randomly. What if we could use pretrained word embeddings to intialize it instead?
Let's take an example: imagine that you have the word pizza in your corpus. Following the previous architecture, you would initialize it to a 300 dimension vector of random float values. This is perfectly fine. You can do that, and this embedding will adjust an evolve throughout training. However, what you could do instead of randomly picking a vector for pizza is using an embedding for this word that has been learnt from another model on a very large corpus. This is a special kind of transfer learning.
Using the knowledge from an external embedding can enhance the precision of your RNN because it integrates new information (lexical and semantic) about the words, an information that has been trained and distilled on a very large corpus of data.
The pretrained embedding we'll be using is GloVe.
Official documentation: GloVe is an unsupervised learning algorithm for obtaining vector representations for words. Training is performed on aggregated global wordword cooccurrence statistics from a corpus, and the resulting representations showcase interesting linear substructures of the word vector space.
The GloVe embeddings I'll be using are trained on a very large common internet crawl that includes:
 840 Billion tokens,
 2.2 million size vocab
The zipped file is 2.03 GB download. Beware, this file cannot be easily loaded on a standard laptop.
The dimension of GloVe embeddings is 300.
GloVe embeddings come in raw text data, where each line contains a word and 300 floats (the corresponding embedding). So the first thing to do is convert this structure to a python dictionary.
def get_coefs(word, *arr):
try:
return word, np.asarray(arr, dtype='float32')
except:
return None, None
embeddings_index = dict(get_coefs(*o.strip().split()) for o
in tqdm_notebook(open('./embeddings/glove.840B.300d.txt')))
embed_size=300
for k in tqdm_notebook(list(embeddings_index.keys())):
v = embeddings_index[k]
try:
if v.shape != (embed_size, ):
embeddings_index.pop(k)
except:
pass
embeddings_index.pop(None)
Once the embedding index in created, we extract all the vectors, we stack them together and compute their mean and standard deviation.
values = list(embeddings_index.values())
all_embs = np.stack(values)
emb_mean, emb_std = all_embs.mean(), all_embs.std()
Now we generate the embedding matrix. We will initialize it following a normal distribution of mean=emb_mean and std=emb_std.
Then we go through the 80000 words of our corpus. For each word, if it is contained in GloVe, we pick its embedding.
Otherwise, we pass.
word_index = tokenizer.word_index
nb_words = MAX_NB_WORDS
embedding_matrix = np.random.normal(emb_mean, emb_std,
(nb_words, embed_size))
oov = 0
for word, i in tqdm_notebook(word_index.items()):
if i >= MAX_NB_WORDS: continue
embedding_vector = embeddings_index.get(word)
if embedding_vector is not None:
embedding_matrix[i] = embedding_vector
else:
oov += 1
print(oov)
def get_rnn_model_with_glove_embeddings():
embedding_dim = 300
inp = Input(shape=(MAX_LENGTH, ))
x = Embedding(MAX_NB_WORDS, embedding_dim, weights=[embedding_matrix],
input_length=MAX_LENGTH, trainable=True)(inp)
x = SpatialDropout1D(0.3)(x)
x = Bidirectional(GRU(100, return_sequences=True))(x)
avg_pool = GlobalAveragePooling1D()(x)
max_pool = GlobalMaxPooling1D()(x)
conc = concatenate([avg_pool, max_pool])
outp = Dense(1, activation="sigmoid")(conc)
model = Model(inputs=inp, outputs=outp)
model.compile(loss='binary_crossentropy',
optimizer='adam',
metrics=['accuracy'])
return model
rnn_model_with_embeddings = get_rnn_model_with_glove_embeddings()
filepath="./models/rnn_with_embeddings/weightsimprovement
{epoch:02d}{val_acc:.4f}.hdf5"
checkpoint = ModelCheckpoint(filepath, monitor='val_acc',
verbose=1, save_best_only=True, mode='max')
batch_size = 256
epochs = 4
history = rnn_model_with_embeddings.fit(x=padded_train_sequences,
y=y_train,
validation_data=(padded_test_sequences, y_test),
batch_size=batch_size,
callbacks=[checkpoint],
epochs=epochs,
verbose=1)
best_rnn_model_with_glove_embeddings =
load_model('./models/rnn_with_embeddings/weightsimprovement030.8372.hdf5')
y_pred_rnn_with_glove_embeddings = best_rnn_model_with_glove_embeddings.predict(
padded_test_sequences, verbose=1, batch_size=2048)
y_pred_rnn_with_glove_embeddings = pd.DataFrame(y_pred_rnn_with_glove_embeddings,
columns=['prediction'])
y_pred_rnn_with_glove_embeddings['prediction'] =
y_pred_rnn_with_glove_embeddings['prediction'].
map(lambda p: 1 if p >= 0.5 else 0)
y_pred_rnn_with_glove_embeddings.to_csv
('./predictions/y_pred_rnn_with_glove_embeddings.csv',
index=False)
y_pred_rnn_with_glove_embeddings = pd.read_csv
('./predictions/y_pred_rnn_with_glove_embeddings.csv')
print(accuracy_score(y_test, y_pred_rnn_with_glove_embeddings))
83.7% accuracy ! Transfer learning from external word embeddings works! For the rest of the tutorial, I'll be using GloVe embeddings in the embedding matrix.
6  Multichannel Convolutional Neural Network
In this section, I'm experimenting a convolutional neural network architecture I read about here. CNNs are generally used in computer vision. However, we've recently started applying them to NLP tasks and the results were promising.
Let's briefly see what happens when we use convnets on text data. To explain this, I'm borrowing this famous diagram (below) from wildm.com (a very good blog!)
Let's consider the example it used: I like this movie very much ! (7 tokens)
 The embedding dimension of each word is 5. Therefore this sentence is represented by a matrix of dimension (7,5). You can think of it as an "image" (~ a matrix of digits/floats).
 6 Filters, 2 of size (2, 5) (3, 5) and (4, 5) are applied on this matrix. The particularity of these filters is that they are not square matrices and their width is equal to the embedding matrix's width. So the result of each convolution will be a column vector.
 Each column vector resulting from the convolution is subsampled using a maxpooling operation.
 The results of the maxpooling operations are concatenated in a final vector that is passed to a softmax function for classification.
What is the intuition behind ?
The result of each convolution will fire when a special pattern is detected. By varying the size of the kernels and concatenating their outputs, you're allowing yourself to detect patterns of multiples sizes (2, 3, or 5 adjacent words).
Patterns could be expressions (word ngrams?) like "I hate", "very good" and therefore CNNs can identify them in the sentence regardless of their position.
def get_cnn_model():
embedding_dim = 300
filter_sizes = [2, 3, 5]
num_filters = 256
drop = 0.3
inputs = Input(shape=(MAX_LENGTH,), dtype='int32')
embedding = Embedding(input_dim=MAX_NB_WORDS,
output_dim=embedding_dim,
weights=[embedding_matrix],
input_length=MAX_LENGTH,
trainable=True)(inputs)
reshape = Reshape((MAX_LENGTH, embedding_dim, 1))(embedding)
conv_0 = Conv2D(num_filters,
kernel_size=(filter_sizes[0], embedding_dim),
padding='valid', kernel_initializer='normal',
activation='relu')(reshape)
conv_1 = Conv2D(num_filters,
kernel_size=(filter_sizes[1], embedding_dim),
padding='valid', kernel_initializer='normal',
activation='relu')(reshape)
conv_2 = Conv2D(num_filters,
kernel_size=(filter_sizes[2], embedding_dim),
padding='valid', kernel_initializer='normal',
activation='relu')(reshape)
maxpool_0 = MaxPool2D(pool_size=(MAX_LENGTH  filter_sizes[0] + 1, 1),
strides=(1,1), padding='valid')(conv_0)
maxpool_1 = MaxPool2D(pool_size=(MAX_LENGTH  filter_sizes[1] + 1, 1),
strides=(1,1), padding='valid')(conv_1)
maxpool_2 = MaxPool2D(pool_size=(MAX_LENGTH  filter_sizes[2] + 1, 1),
strides=(1,1), padding='valid')(conv_2)
concatenated_tensor = Concatenate(axis=1)(
[maxpool_0, maxpool_1, maxpool_2])
flatten = Flatten()(concatenated_tensor)
dropout = Dropout(drop)(flatten)
output = Dense(units=1, activation='sigmoid')(dropout)
model = Model(inputs=inputs, outputs=output)
adam = Adam(lr=1e4, beta_1=0.9, beta_2=0.999, epsilon=1e08, decay=0.0)
model.compile(optimizer=adam, loss='binary_crossentropy',
metrics=['accuracy'])
return model
cnn_model_multi_channel = get_cnn_model()
plot_model(cnn_model_multi_channel,
to_file='./images/article_5/cnn_model_multi_channel.png',
show_shapes=True,
show_layer_names=True)
filepath="./models/cnn_multi_channel/weightsimprovement
{epoch:02d}{val_acc:.4f}.hdf5"
checkpoint = ModelCheckpoint(filepath, monitor='val_acc',
verbose=1, save_best_only=True, mode='max')
batch_size = 256
epochs = 4
history = cnn_model_multi_channel.fit(x=padded_train_sequences,
y=y_train,
validation_data=(padded_test_sequences, y_test),
batch_size=batch_size,
callbacks=[checkpoint],
epochs=epochs,
verbose=1)
best_cnn_model = load_model
('./models/cnn_multi_channel/weightsimprovement040.8264.hdf5')
y_pred_cnn_multi_channel = best_cnn_model.predict(padded_test_sequences,
verbose=1, batch_size=2048)
y_pred_cnn_multi_channel = pd.DataFrame(y_pred_cnn_multi_channel,
columns=['prediction'])
y_pred_cnn_multi_channel['prediction'] =
y_pred_cnn_multi_channel['prediction'].map(lambda p: 1 if p >= 0.5 else 0)
y_pred_cnn_multi_channel.to_csv('./predictions/y_pred_cnn_multi_channel.csv',
index=False)
y_pred_cnn_multi_channel = pd.read_csv
('./predictions/y_pred_cnn_multi_channel.csv')
print(accuracy_score(y_test, y_pred_cnn_multi_channel))
82.6% accuracy , we're less precise than RNNs but still better than BOW models. Maybe an investigation of the hyperparameters (number of filters, and size) gives an edge?
7  Recurrent + Convolutional neural network
RNNs are powerful. Howerer, some people found out that they could maek them more robust by adding a convolutional layer on top of the reccurrent layer.
The rational behind is that RNNs allow you to embed the information about the sequence and previous words and CNN takes this embedding and extract local features from it. Having these two layers working together is a winning combination.
More about this here.
def get_rnn_cnn_model():
embedding_dim = 300
inp = Input(shape=(MAX_LENGTH, ))
x = Embedding(MAX_NB_WORDS, embedding_dim, weights=[embedding_matrix],
input_length=MAX_LENGTH, trainable=True)(inp)
x = SpatialDropout1D(0.3)(x)
x = Bidirectional(GRU(100, return_sequences=True))(x)
x = Conv1D(64, kernel_size = 2, padding = "valid",
kernel_initializer = "he_uniform")(x)
avg_pool = GlobalAveragePooling1D()(x)
max_pool = GlobalMaxPooling1D()(x)
conc = concatenate([avg_pool, max_pool])
outp = Dense(1, activation="sigmoid")(conc)
model = Model(inputs=inp, outputs=outp)
model.compile(loss='binary_crossentropy',
optimizer='adam',
metrics=['accuracy'])
return model
rnn_cnn_model = get_rnn_cnn_model()
plot_model(rnn_cnn_model, to_file='./images/article_5/rnn_cnn_model.png',
show_shapes=True, show_layer_names=True)
filepath="./models/rnn_cnn/weightsimprovement
{epoch:02d}{val_acc:.4f}.hdf5"
checkpoint = ModelCheckpoint(filepath, monitor='val_acc',
verbose=1,
save_best_only=True, mode='max')
batch_size = 256
epochs = 4
history = rnn_cnn_model.fit(x=padded_train_sequences,
y=y_train,
validation_data=(padded_test_sequences, y_test),
batch_size=batch_size,
callbacks=[checkpoint],
epochs=epochs,
verbose=1)
best_rnn_cnn_model = load_model('./models/rnn_cnn/weights
improvement030.8379.hdf5')
y_pred_rnn_cnn = best_rnn_cnn_model.predict(padded_test_sequences,
verbose=1, batch_size=2048)
y_pred_rnn_cnn = pd.DataFrame(y_pred_rnn_cnn, columns=['prediction'])
y_pred_rnn_cnn['prediction'] = y_pred_rnn_cnn['prediction']
.map(lambda p: 1 if p >= 0.5 else 0)
y_pred_rnn_cnn.to_csv('./predictions/y_pred_rnn_cnn.csv', index=False)
y_pred_rnn_cnn = pd.read_csv('./predictions/y_pred_rnn_cnn.csv')
print(accuracy_score(y_test, y_pred_rnn_cnn))
83.8% accuracy . Best model so far.
8  Conclusion
We've run seven different models. Let's see how they compare:
import seaborn as sns
from sklearn.metrics import roc_auc_score
sns.set_style("whitegrid")
sns.set_palette("pastel")
predictions_files = os.listdir('./predictions/')
predictions_dfs = []
for f in predictions_files:
aux = pd.read_csv('./predictions/{0}'.format(f))
aux.columns = [f.strip('.csv')]
predictions_dfs.append(aux)
predictions = pd.concat(predictions_dfs, axis=1)
scores = {}
for column in tqdm_notebook(predictions.columns, leave=False):
if column != 'y_true':
s = accuracy_score(predictions['y_true'].values,
predictions[column].values)
scores[column] = s
scores = pd.DataFrame([scores], index=['accuracy'])
mapping_name = dict(zip(list(scores.columns),
['Char ngram + LR', '(Word + Char ngram) + LR',
'Word ngram + LR', 'CNN (multi channel)',
'RNN + CNN', 'RNN no embd.', 'RNN +
GloVe embds.']))
scores = scores.rename(columns=mapping_name)
scores = scores[['Word ngram + LR', 'Char ngram + LR', '
(Word + Char ngram) + LR',
'RNN no embd.', 'RNN + GloVe embds.', '
CNN (multi channel)',
'RNN + CNN']]
scores = scores.T
ax = scores['accuracy'].plot(kind='bar',
figsize=(16, 5),
ylim=(scores.accuracy.min()*0.97,
scores.accuracy.max() * 1.01),
color='red',
alpha=0.75,
rot=45,
fontsize=13)
ax.set_title('Comparative accuracy of the different models')
for i in ax.patches:
ax.annotate(str(round(i.get_height(), 3)),
(i.get_x() + 0.1, i.get_height() * 1.002),
color='dimgrey', fontsize=14)
Let's quickly check the correlations between the predictions of the models.
fig = plt.figure(figsize=(10, 5))
sns.heatmap(predictions.drop('y_true', axis=1).corr(method='kendall'),
cmap="Blues", annot=True);
Conclusion
Here are rapid findings I think are worth sharing:
 Bag of word models using character ngrams can be very efficient. Do not underestimate them!. They are relatively cheap to compute, and also easy to interpret.
 RNNs are powerful. However, you can sometimes pump them with external pretrained embeddings like GloVe. You can also use other popular embeddings such as word2vec and FastText.
 CNNs can be applied to text. They have the main advantage of being very fast to train. Besides, their ability to extract local features out of text is particularly interesting to nlp tasks.
 RNNs and CNNs can be stacked together to take advantags of both architectures.
This post was quite long, I hope you've enjoyed it. Don't hesitate to comment if you have any question or recommendation.
Some helpful links to explore
Here are great resources I used when writing this post:
Bio: Ahmed Besbes is Senior Data Scientist at Capgemini Consulting, based in Paris, France. He is a Deep learning practitioner, blogger and opensource aficionado. Handson, obsessed with neat design and automation. Blog: http://ahmedbesbes.com ;
Github: https://github.com/ahmedbesbes ; Kaggle profile: https://www.kaggle.com/ahmedbesbes
Original. Reposted with permission.
Related:
Top Stories Past 30 Days

