Join us and get access to hundreds of tutorials and a community of expert Pythonistas.

Unlock This Lesson

This lesson is for members only. Join us and get access to hundreds of tutorials and a community of expert Pythonistas.

Unlock This Lesson

Hint: You can adjust the default video playback speed in your account settings.
Hint: You can set the default subtitles language in your account settings.
Sorry! Looks like there’s an issue with video playback 🙁 This might be due to a temporary outage or because of a configuration issue with your browser. Please see our video player troubleshooting guide to resolve the issue.

Using Pretrained Word Embeddings

Here are resources for more information about Word2Vec and GloVe:

00:00 Some popular pretrained embeddings include the Word2Vec from Google and the GloVe from the NLP team at Stanford University. In this course, you’ll use GloVe for its size and speed. Word2Vec is larger, but it is also more accurate, so you can try it once you’ve seen GloVe in your code.

00:20 The next snippet will download the GloVe data set and extract it using utility methods from earlier in the course. It’s trained on 6 billion words, so the file size is over 800 megabytes and will take a while to process.

00:33 I’ll speed it up through the magic of video. The file contains a list of words with the embedding vector for that word. Use the next code to get a reduced version of the embedding matrix.

00:46 The embedding matrix is stored in an array with 1,747 rows, which is the length of the vocabulary, and 50 columns, which is the size of the embedding.

00:57 Before using this matrix in a model, take a look at the number of non-zero elements it contains. There’s a little more than 95% of the vocabulary that is contained by the matrix. To use the matrix in the model, assign the embedding matrix to the weights keyword argument of the layer. Also, set the trainable keyword argument to False. The matrix has already been trained, so you don’t need to do anything else for it to work. Train the model, test it, and look at the results. And this is better.

01:33 Can you get more by training the embedding layer more? Let’s see. Set the trainable keyword argument to True, then train and test the model once more and plot the results. You’re getting there!

01:49 Again, this model needs to be trained no more than 20 epochs.

01:54 In the next video, you’ll see an advanced type of neural network which can yield even better inferences.

Become a Member to join the conversation.