A Guide to Feature Engineering in NLP

Overview Feature engineering in NLP is understanding the context of the text. In this blog, we will look at some of the common feature engineering in NLP. We will compare the results of a classification task with and without doing feature engineering   Table of Content Introduction NLP task overview List of features with code Implementation Results comparison with and without doing feature engineering Conclusion Introduction   “If 80 percent of our work is data preparation, then ensuring data quality […]

Read more

Language Detection Using Natural Language Processing

Introduction Every Machine Learning enthusiast has a dream of building/working on a cool project, isn’t it? Mere understandings of the theory aren’t enough, you need to work on projects, try to deploy them, and learn from them. Moreover, working on specific domains like NLP gives you wide opportunities and problem statements to explore. Through this article, I wish to introduce you to an amazing project, the Language Detection model using Natural Language Processing. This will take you through a real-world […]

Read more

A Hands-On Introduction to Hugging Face’s AutoNLP 101

Hugging Face, founded in 2016, has revolutionized the way people approach Natural Language Processing in this day and age. Based in New York, Hugging Face started out as a chatbot company with its primary focus today on the Transformers library and helping the developers integrate NLP into their own products or services. Hugging Face has made it incredibly easy for an individual to train their data on huge state-of-the-art models only with a couple of lines. Solving NLP, one commit […]

Read more

Sentiment Analysis: VADER or TextBlob?

This article was published as a part of the Data Science Blogathon. What Is Sentiment Analysis? Conclusions are integral to practically all human exercises and are key influencers of our practices. Our convictions and impression of the real world, and the decisions we make, are, to an impressive degree, molded upon how others see and assess the world. Therefore, when we have to settle on a choice, we regularly search out the assessments of others. Opinions and their related concepts […]

Read more

Sentiment Analysis: Predicting Sentiment Of COVID-19 Tweets

This article was published as a part of the Data Science Blogathon. Introduction Hi folks, I hope you are doing well in these difficult times! We all are going through the unprecedented time of the Corona Virus pandemic. Some people lost their lives, but many of us successfully defeated this new strain i.e. Covid-19. The virus was declared a pandemic by World Health Organization on 11th March 2020. This article will analyze various types of “Tweets” gathered during pandemic times. […]

Read more

Introduction to Hugging Face’s Transformers v4.3.0 and its First Automatic Speech Recognition Model – Wav2Vec2

Overview Hugging Face has released Transformers v4.3.0 and it introduces the first Automatic Speech Recognition model to the library: Wav2Vec2 Using one hour of labeled data, Wav2Vec2 outperforms the previous state of the art on the 100-hour subset while using 100 times less labeled data Using just ten minutes of labeled data and pre-training on 53k hours of unlabeled data Wav2Vec2 achieves 4.8/8.2 WER Understand Wav2Vec2 implementation using transformers library on audio to text generation   Introduction Transformers has been […]

Read more

Streamlit Web API for NLP: Tweet Sentiment Analysis

This article was published as a part of the Data Science Blogathon. Introduction Developing Web Apps for data models has always been a hectic task for non-web developers. For developing Web API we need to make the front end as well as back end platform. That is not an easy task. But then python comes to the rescue with its very fascinating frameworks like Streamlit, Flassger, FastAPI. These frameworks help us to build web APIs very elegantly, without worrying about […]

Read more

Emotion classification on Twitter Data Using Transformers

Introduction The world of Natural language processing is recently overtaken by the invention of Transformers. Transformers are entirely indifferent to the conventional sequence-based networks. RNNs are the initial weapon used for sequence-based tasks like text generation, text classification, etc. But with the arrival of LSTM and GRU cells, the issue with capturing long-term dependency in the text got resolved. But learning the model with LSTM cells is a hard task as we cannot make it learn parallelly.

Read more

Hands-On Tutorial on Stack Overflow Question Tagging

This article was published as a part of the Data Science Blogathon. Background I won’t be lying if I assert that every developer/engineer/student has used the website Stack Overflow more than once in their journey. Widely considered as one of the largest and more trusted websites for developers to learn and share their knowledge, the website presently hosts in excess of 10,000,000 questions. In this post, we try to predict the question tags based on the question text asked on […]

Read more

Step by step guide to building sentiment analysis model using graphlab

I have been using graph lab for quite some time now. The first Kaggle competition I used it for was Click Trough Rate (CTR) and I was amazed to see the speed at which it can crunch such big data. Over last few months, I have realised much broader applications of GraphLab. In this article I will take up the text mining capability of GraphLab and solve one of the Kaggle problems. I will be referring to this problem with […]

Read more
1 2 3 4 6