Mastering Next Word Prediction with BI-LSTM: A Comprehensive Guide
Identifying the following word is the task of next-word prediction, also known as language modeling. One of the NLP‘s benchmark tasks is language modeling. In its most basic form, it entails picking the word that follows a string of words based on them that is most likely to occur. In many different fields, language modeling has a wide variety of applications.
- Recognize the underlying ideas and principles behind the numerous models used in statistical analysis, machine learning, and data science.
- Learn how to create predictive models, including regression, classification, clustering, etc., to generate precise predictions and types based on data.
- Understand the principles of overfitting and underfitting, and learn how to evaluate model performance using measures like accuracy, precision, recall, etc.
- Learn how to preprocess data and identify pertinent characteristics for modeling.
- Learn how to tweak hyperparameters and optimize models using grid search and cross-validation.
This article was published as a part of the Data Science Blogathon.
Table of contents
- Applications of Language Modeling
- Import Necessary Libraries and Packages
- Dataset Information
- Display Titles of Various Articles and Preprocess Them
- Prepare Features and Labels
- The Architecture of Bidirectional LSTM Neural Network
- Bi-LSTM Neural Network Model training
- Plotting Model Accuracy and Loss
- Predicting the Next Word of the Title
- Frequently Asked Questions
Applications of Language Modeling
Here are some notable applications of language modeling:
Mobile Keyboard Text Recommendation
A function on smartphone keyboards called mobile keyboard text recommendation, or predictive text or auto-suggestions, suggests words or phrases as you write. It seeks to make typing faster and less error-prone and to offer more precise and contextually appropriate recommendations.
Google Search Auto-Completion
Every time we use a search engine like Google to look for anything, we receive many ideas, and as we keep adding phrases, the recommendations grow better and more relevant to our current search. How will it happen, then?
Natural language processing (NLP) technology makes it feasible. Here, we’ll employ natural language processing (NLP) to create a prediction model utilizing a bidirectional LSTM (Long short-term memory) model to foretell the sentence’s remaining words.
Import Necessary Libraries and Packages
Importing the necessary libraries and packages to construct a next-word prediction model using a bidirectional LSTM would be best. A sample of the libraries you’ll generally require is shown below:
import pandas as pd import os import numpy as np import tensorflow as tf from tensorflow.keras.preprocessing.sequence import pad_sequences from tensorflow.keras.layers import Embedding, LSTM, Dense, Bidirectional from tensorflow.keras.preprocessing.text import Tokenizer from tensorflow.keras.models import Sequential from tensorflow.keras.optimizers import Adam
Understanding the features and attributes of the dataset you’re dealing with requires knowledge. The following seven publications’ medium articles, selected at random and published in 2019, are included in this dataset:
- Towards Data Science
- UX Collective
- The Startup
- The Writing Cooperative
- Data Driven Investor
- Better Humans
- Better Marketing
medium_data = pd.read_csv('../input/medium-articles-dataset/medium_data.csv') medium_data.head()
Here, we have ten different fields and 6508 records but we will only use the title field for predicting the next word.
print("Number of records: ", medium_data.shape) print("Number of fields: ", medium_data.shape)
By looking through and comprehending the dataset information, you may choose the preprocessing procedures, model, and evaluation metrics for your next word prediction challenge.
Display Titles of Various Articles and Preprocess Them
Let’s have a look at a few sample titles to illustrate the preparation of article titles:
Removing Unwanted Characters and Words in Titles
Preprocessing text data for prediction tasks sometimes includes removing undesirable letters and phrases from titles. Unwanted letters and words might contaminate the data with noise and add needless complexity, thereby lowering the model’s performance and accuracy.
- Unwanted Characters:
- Punctuation: You should remove exclamation points, question marks, commas, and other punctuation. Typically, you can safely discard them because they usually don’t help with the prediction assignment
- Special Characters: Remove non-alphanumeric symbols, such as dollar signs, @ symbols, hashtags, and other special characters, that are unnecessary for the prediction job.
- HTML Tags: If the titles have HTML markups or tags, remove them using the proper tools or libraries to extract the text.
- Unwanted Words:
- Stop Words: Remove common stop words such as “a,” “an,” “the,” “is,” “in,” and other frequently occurring words that do not carry significant meaning or predictive power.
- Irrelevant Words: Identify and remove specific words that are not relevant to the prediction task or domain. For example, if you are predicting movie genres, words like “movie” or “film” may not provide helpful information.
medium_data['title'] = medium_data['title'].apply(lambda x: x.replace(u'\xa0',u' ')) medium_data['title'] = medium_data['title'].apply(lambda x: x.replace('\u200a',' '))
Tokenization divides the text into tokens, words, subwords, or characters and then assigns a unique ID or index to each token, creating a word index or Vocabulary.
The tokenization process involves the following steps:
Text preprocessing: Preprocess the text by eliminating punctuation, changing it to lowercase, and taking care of any particular task- or domain-specific needs.
Tokenization: Dividing the preprocessed text into separate tokens by predetermined rules or methods. Regular expressions, separating by whitespace, and employing specialized tokenizers are all common tokenization techniques.
Increasing Vocabulary You can make a dictionary, also called a word index, by assigning each token a unique ID or index. In this process, each ticket is mapped to the relevant index value.
tokenizer = Tokenizer(oov_token='<oov>') # For those words which are not found in word_index tokenizer.fit_on_texts(medium_data['title']) total_words = len(tokenizer.word_index) + 1 print("Total number of words: ", total_words) print("Word: ID") print("------------") print("<oov>: ", tokenizer.word_index['<oov>']) print("Strong: ", tokenizer.word_index['strong']) print("And: ", tokenizer.word_index['and']) print("Consumption: ", tokenizer.word_index['consumption'])
By transforming text into a vocabulary or word index, you can create a lookup table representing the text as a collection of numerical indexes. Each unique word in the text receives a corresponding index value, allowing for further processing or modeling operations that require numerical input.
Titles Text into Sequences and Make N_gram Model.
These stages can be used to build an n-gram model for accurate prediction based on title sequences:
- Convert Titles to Sequences: Use a tokenizer to turn each title into a string of tokens or manually separate each slip into its constituent words. Assign each word in the lexicon a distinct number index.
- Generate n-grams: From the sequences, make n-grams. A continuous run of n-title tokens is called an n-gram.
- Count the Frequency: Determine the frequency at which each n-gram appears in the dataset.
- Build the n-gram Model: Create the n-gram model using the n-gram frequencies. The model keeps track of each token probability given the previous n-1 tokens. This can be displayed as a lookup table or a dictionary.
- Predict the Next Word: The expected next token in an n-1-token sequence may be identified using the n-gram model. To do this, it is necessary to find the probability in the algorithm and select a token with the greatest likelihood.
You can use these stages to build an n-gram model that utilizes the titles’ sequences to predict the next word or token. Based on the training data, this method can produce accurate predictions since it captures the statistical relationships and trends in the language usage of the titles.
input_sequences =  for line in medium_data['title']: token_list = tokenizer.texts_to_sequences([line]) #print(token_list) for i in range(1, len(token_list)): n_gram_sequence = token_list[:i+1] input_sequences.append(n_gram_sequence) # print(input_sequences) print("Total input sequences: ", len(input_sequences))
Make All Titles the Same Length by Using Padding
You may use padding to ensure that each title is the same size by following these steps:
- Find the longest title in your dataset by comparing all the other titles.
- Repeat this process for each title, comparing each one’s length to the overall limit.
- When a title is too short, it should be extended using a specific padding token or character.
- For each title in your dataset, carry out the padding procedure again.
Padding will ensure that all titles are the same length and will provide consistency for post-processing or model training.
# pad sequences max_sequence_len = max([len(x) for x in input_sequences]) input_sequences = np.array(pad_sequences(input_sequences, maxlen=max_sequence_len, padding='pre')) input_sequences
Prepare Features and Labels
In the given scenario, if we consider the last element of each input sequence as the label, we can perform one-hot encoding on the titles to represent them as vectors corresponding to the total number of unique words.
# create features and label xs, labels = input_sequences[:,:-1],input_sequences[:,-1] ys = tf.keras.utils.to_categorical(labels, num_classes=total_words) print(xs) print(labels) print(ys)
The Architecture of Bidirectional LSTM Neural Network
Recurrent neural networks (RNNs) with Long Short-Term Memory (LSTM) can collect and hold information across extensive sequences. LSTM networks use specialized memory cells and gating techniques to overcome the constraints of regular RNNs, which frequently struggle with the vanishing gradient problem and have trouble maintaining long-term dependence.
The critical feature of LSTM networks is the cell state, which serves as a memory unit that can store information over time. The cell state is protected and controlled by three main gates: the forget gate, the input gate, and the output gate. These gates regulate the flow of information into, out of, and within the LSTM cell, allowing the network to remember or forget information at different time steps selectively.
Learn More: Long Short Term Memory | Architecture Of LSTM
Bi-LSTM Neural Network Model training
Numerous crucial procedures must be followed while training a bidirectional LSTM (Bi-LSTM) neural network model. The first step is compiling a training dataset with input and output sequences corresponding to them, indicating the next word. The text data must be preprocessed by being divided into separate lines, removing the punctuation, and changing the case to lowercase.
model = Sequential() model.add(Embedding(total_words, 100, input_length=max_sequence_len-1)) model.add(Bidirectional(LSTM(150))) model.add(Dense(total_words, activation='softmax')) adam = Adam(lr=0.01) model.compile(loss='categorical_crossentropy', optimizer=adam, metrics=['accuracy']) history = model.fit(xs, ys, epochs=50, verbose=1) #print model.summary() print(model)
By calling the fit() method, the model is trained. The training data consists of the input sequences (xs) and matching output sequences (ys). The model proceeds through 50 iterations, going through the whole training set. During the training process, the training progress is shown (verbose=1).
Plotting Model Accuracy and Loss
Plotting a model’s accuracy and loss throughout training offers insightful information about how well it performs and how training is going. The mistake or disparity between the anticipated and actual values is called loss. Whereas the percentage of accurate predictions generated by the model is known as accuracy.
import matplotlib.pyplot as plt def plot_graphs(history, string): plt.plot(history.history[string]) plt.xlabel("Epochs") plt.ylabel(string) plt.show() plot_graphs(history, 'accuracy')
Predicting the Next Word of the Title
A fascinating challenge in natural language processing is guessing the following word in a title. Models can propose the most likely talk by looking for patterns and correlations in text data. This predictive power makes applications like text suggestion systems and autocomplete possible. Sophisticated approaches like RNNs and transformer-based architectures increase accuracy and capture contextual relationships.
seed_text = "implementation of" next_words = 2 for _ in range(next_words): token_list = tokenizer.texts_to_sequences([seed_text]) token_list = pad_sequences([token_list], maxlen=max_sequence_len-1, padding='pre') predicted = model.predict_classes(token_list, verbose=0) output_word = "" for word, index in tokenizer.word_index.items(): if index == predicted: output_word = word break seed_text += " " + output_word print(seed_text)
In conclusion, training a model to predict the subsequent word in a string of words is the exciting natural language processing challenge known as next-word prediction using a Bidirectional LSTM. Here’s the conclusion summarized in bullet points:
- The potent deep learning architecture BI-LSTM for sequential data processing may capture long-range relationships and phrase context.
- To prepare raw text data for BI-LSTM training, data preparation is essential. This includes tokenization, vocabulary generation, and text vectorization.
- Creating a loss function, building the model using an optimizer, fitting it to preprocessed data, and assessing its performance on validation sets are the steps in training the BI-LSTM model.
- BI-LSTM next word prediction takes a combination of theoretical knowledge and hands-on experimentation to master.
- Auto-completion, language creation, and text suggestion algorithms are examples of next-word prediction model applications.
Applications for next-word prediction include chatbots, machine translation, and text completion. You can create more precise and context-aware next-word prediction models with more research and improvement.
Frequently Asked Questions
A. Next word prediction is a NLP task where a model predicts the most likely word to follow a given sequence of words or context. It aims to generate coherent and contextually relevant suggestions for the next word based on the patterns and relationships learned from training data.
A. Next-word prediction commonly uses Recurrent Neural Networks (RNNs) and their variants, such as Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU). Additionally, models like Transformer-based architectures, such as the GPT (Generative Pre-trained Transformer) models, have also shown significant advancements in this task.
A. Typically, when preparing training data for next-word prediction, you split text into sequences of words and create input-output pairs. The corresponding output represents the following word in the text for each input sequence. Preprocessing the text involves removing punctuation, converting words to lowercase, and tokenizing the text into individual words.
A. You can evaluate the performance of a next-word prediction model using evaluation metrics such as perplexity, accuracy, or top-k accuracy. Perplexity measures how well the model predicts the next word given the context. Accuracy metrics compare the predicted word with the ground truth, while top-k accuracy considers the model’s prediction within the top-k most probable comments.
The media shown in this article is not owned by Analytics Vidhya and is used at the Author’s discretion.