Word2Vec For Word Embeddings -A Beginner’s Guide

This article was published as a part of the Data Science Blogathon Why are word embeddings needed? Let us consider the two sentences – “You can scale your business.” and “You can grow your business.”. These two sentences have the same meaning. If we consider a vocabulary considering these two sentences, it will constitute of these words: {You, can, scale, grow, your, business}. A one-hot encoding of these words would create a vector of length 6. The encodings for each of […]

Read more

Text Classification & Word Representations using FastText (An NLP library by Facebook)

Introduction If you put a status update on Facebook about purchasing a car -don’t be surprised if Facebook serves you a car ad on your screen. This is not black magic! This is Facebook leveraging the text data to serve you better ads. The picture below takes a jibe at a challenge while dealing with text data. Well, it clearly failed in the above attempt to deliver the right ad. It is all the more important to capture the context […]

Read more

An Intuitive Understanding of Word Embeddings: From Count Vectors to Word2Vec

Introduction Before we start, have a look at the below examples. You open Google and search for a news article on the ongoing Champions trophy and get hundreds of search results in return about it. Nate silver analysed millions of tweets and correctly predicted the results of 49 out of 50 states in 2008 U.S Presidential Elections. You type a sentence in google translate in English and get an Equivalent Chinese conversion.   So what do the above examples have […]

Read more

An Introductory Guide to Understand how ANNs Conceptualize New Ideas (using Embedding)

Introduction Here’s something you don’t hear everyday – everything we perceive is just a best case probabilistic prediction by our brain, based on our past encounters and knowledge gained through other mediums. This might sound extremely counter intuitive because we have always imagined that our brain mostly gives us deterministic answers. We’ll do a small experiment to showcase this logic. Take a look at the below image: Q1. Do you see a human ? Q2. Can you identify the person? […]

Read more

Building a Recommendation System using Word2vec: A Unique Tutorial with Case Study in Python

Overview Recommendation engines are ubiquitous nowadays and data scientists are expected to know how to build one Word2vec is an ultra-popular word embeddings used for performing a variety of NLP tasks We will use word2vec to build our own recommendation system. Curious how NLP and recommendation engines combine? Let’s find out!   Introduction Be honest – how many times have you used the ‘Recommended for you’ section on Amazon? Ever since I found out a few years back that machine […]

Read more

An Essential Guide to Pretrained Word Embeddings for NLP Practitioners

Overview Understand the importance of pretrained word embeddings Learn about the two popular types of pretrained word embeddings – Word2Vec and GloVe Compare the performance of pretrained word embeddings and learning embeddings from scratch   Introduction How do we make machines understand text data? We know that machines are supremely adept at dealing and working with numerical data but they become sputtering instruments if we feed raw text data to them. The idea is to create a representation of words […]

Read more

Must-Read Tutorial to Learn Sequence Modeling (deeplearning.ai Course #5)

Introduction The ability to predict what comes next in a sequence is fascinating. It’s one of the reasons I became interested in data science! Interestingly – human mind is really good at it, but that is not the case with machines. Given a mysterious plot in a book, the human brain will start creating outcomes. But, how to teach machines to do something similar? Thanks to Deep Learning – we can do lot more today than what was possible a […]

Read more

Information Retrieval using word2vec based Vector Space Model

Overview Learn about Information Retrieval (IR), Vector Space Models (VSM), and Mean Average Precision (MAP) Create a project on Information Retrieval using word2vec based Vector Space Model   Introduction “Google it!”- Isn’t it something we say every day? Whenever we come across something that we don’t know about, we “Google it.” Google Search is a great tool that can be used for even finding a needle from a haystack. This generation absolutely relies on Google for answers to all kinds […]

Read more