This page contains the guided laboratory of the Embedding Spaces topic for the Deep Learning course at the Master in Artificial Inteligence of the Universitat Politècnica de Catalunya.
The codes introduced in this guided lab can be found here
Some necessary resources are found within gpfs. This must be copied to your local .keras/models folder. Their location is: /gpfs/scratch/bsc28/hpai/storage/data/dl-labs/
Other sources for experimentation
Beyond the codes explained in class, there are other online resources of interest that may be used for experimentation.
The original code of word2vec, as released by its authors, can be found here.
Kaggle word2vec tutorial
Bag of Words Meets Bags of Popcorn is a tutorial for understanding and operating with the word2vec model.
McCormick inspect word2vec
This repository uses gensim in Python to load word2vec pre-trained model, and inspects some of the details of the vocabulary.
word2vec tutorial in TensorFlow
Tutorial of word2vec using TensorFlow: Vector Representations of Words
Word2vec in Keras
Using Gensim Word2Vec Embeddings in Keras. A short post and script regarding using Gensim Word2Vec embeddings in Keras, with example code.
GitHub of Jamie Ryan Kiros visual-semantic-embedding Implementation of the image-sentence embedding method described in “Unifying Visual-Semantic Embeddings with Multimodal Neural Language Models” multimodal-neural-language-models A bare-bones NumPy implementation of “Multimodal Neural Language Models” (Kiros et al, ICML 2014) skip-thoughts Sent2Vec encoder and training code from the paper “Skip-Thought Vectors” neural-storyteller A recurrent neural network for generating little stories about images
Pinterest Multimodal Dataset ToolBox This is a toolbox to download and manage the released part of the Pinterest40M multimodal dataset introduced in the paper Training and Evaluating Multimodal Word Embeddings with Large-scale Web Annotated Images. More information can be found on the Project Page.