How do Transformers Work in NLP? A Guide to the Latest State-of-the-Art Models

Overview The Transformer model in NLP has truly changed the way we work with text data Transformer is behind the recent NLP developments, including Google’s BERT Learn how the Transformer idea works, how it’s related to language modeling, sequence-to-sequence modeling, and how it enables Google’s BERT model   Introduction I love being a data scientist working in Natural Language Processing (NLP) right now. The breakthroughs and developments are occurring at an unprecedented pace. From the super-efficient ULMFiT framework to Google’s […]

Read more

Innoplexus Sentiment Analysis Hackathon: Top 3 Out-of-the-Box Winning Approaches

Overview Hackathons are a wonderful opportunity to gauge your data science knowledge and compete to win lucrative prizes and job opportunities Here are the top 3 approaches from the Innoplexus Sentiment Analysis Hackathon – a superb NLP challenge   Introduction I’m a big fan of hackathons. I’ve learned so much about data science from participating in these hackathons in the past few years. I’ll admit it – I have gained a lot of knowledge through this medium and this, in […]

Read more

How to use a Machine Learning Model to Make Predictions on Streaming Data using PySpark

Overview Streaming data is a thriving concept in the machine learning space Learn how to use a machine learning model (such as logistic regression) to make predictions on streaming data using PySpark We’ll cover the basics of Streaming Data and Spark Streaming, and then dive into the implementation part   Introduction Picture this – every second, more than 8,500 Tweets are sent, more than 900 photos are uploaded on Instagram, more than 4,200 Skype calls are made, more than 78,000 […]

Read more

A Beginner’s Guide to Exploratory Data Analysis (EDA) on Text Data (Amazon Case Study)

The Importance of Exploratory Data Analysis (EDA) There are no shortcuts in a machine learning project lifecycle. We can’t simply skip to the model building stage after gathering the data. We need to plan our approach in a structured manner and the exploratory data analytics (EDA) stage plays a huge part in that. I can say this with the benefit of hindsight having personally gone through this situation plenty of times. In my early days in this field, I couldn’t […]

Read more

People to Follow in the field of Natural Language Processing (NLP)

Overview Text analytics is becoming easier with many people working day and night on each aspect of Natural Language Processing We list a set of people to follow in the field NLP Feel we should include anyone else? Let us know!   Introduction Natural Language Processing has made unstructured text data analysis simpler. With numerous applications, NLP is affecting and adding values to millions of lives. But the problem NLP practitioners face is catching up with the changes that happen […]

Read more

Summarize Twitter Live data using Pretrained NLP models

Introduction Twitter users spend an average of 4 minutes on social media Twitter. On an average of 1 minute, they read the same stuff. It shows that users spend around 25% of their time reading the same stuff. Also, most of the tweets will not appear on your dashboard. You may get to know the trending topics, but you miss not trending topics. In trending topics, you might only read the top 5 tweets and their comments. So, what are […]

Read more

Tired of Reading Long Articles? Text Summarization will make your task easier!

This article was published as a part of the Data Science Blogathon. Introduction Millions of web pages and websites exist on the Internet today. Going through a vast amount of content becomes very difficult to extract information on a certain topic. Google will filter the search results and give you the top ten search results, but often you are unable to find the right content that you need. There is a lot of redundant and overlapping data in the articles […]

Read more

Learning sparse codes from compressed representations with biologically plausible local wiring constraints

Sparse coding is an important method for unsupervised learning of task-independent features in theoretical neuroscience models of neural coding. While a number of algorithms exist to learn these representations from the statistics of a dataset, they largely ignore the information bottlenecks present in fiber pathways connecting cortical areas… For example, the visual pathway has many fewer neurons transmitting visual information to cortex than the number of photoreceptors. Both empirical and analytic results have recently shown that sparse representations can be […]

Read more

Practical Low-Rank Communication Compression in Decentralized Deep Learning

Lossy gradient compression has become a practical tool to overcome the communication bottleneck in centrally coordinated distributed training of machine learning models. However, algorithms for decentralized training with compressed communication over arbitrary connected networks have been more complicated, requiring additional memory and hyperparameters… We introduce a simple algorithm that directly compresses the model differences between neighboring workers using low-rank linear compressors. We prove that our method does not require any additional hyperparameters, converges faster than prior methods, and is asymptotically […]

Read more

Inverting Gradients – How easy is it to break privacy in federated learning?

The idea of federated learning is to collaboratively train a neural network on a server. Each user receives the current weights of the network and in turns sends parameter updates (gradients) based on local data… This protocol has been designed not only to train neural networks data-efficiently, but also to provide privacy benefits for users, as their input data remains on device and only parameter gradients are shared. But how secure is sharing parameter gradients? Previous attacks have provided a […]

Read more
1 700 701 702 703 704 907