# Introduction to Data Imputation

Shashank Singhal 10 Jun, 2024

## Introduction

In data science, missing data can significantly impact analysis quality. This article, part of the Data Science Blogathon, explores data imputation, a technique to replace missing values and retain dataset integrity. Imputation prevents issues like library incompatibility, dataset distortion, and model bias. Weâ€™ll cover key imputation methodsâ€”Complete Case Analysis (CCA), Arbitrary Value Imputation, and Frequent Category Imputationâ€”highlighting their assumptions, benefits, and drawbacks. This guide aims to help you manage missing data effectively, ensuring robust and accurate analyses. Letâ€™s dive into mastering data imputation techniques.

#### Learning Objectives

• Understand the significance of data imputation in data science and its impact on analysis quality.
• Learn about various data imputation methods, including Complete Case Analysis (CCA), Arbitrary Value Imputation, and Frequent Category Imputation.
• Identify the assumptions, benefits, and limitations of different imputation techniques.
• Gain practical skills in applying imputation methods to manage missing data effectively.
• Ensure robust and accurate analyses by mitigating issues related to missing data.

This article was published as a part of theÂ Data Science Blogathon.

## What is Data Imputation?

Imputation is a technique used for replacing the missing data with some substitute value to retain most of the data/information of the dataset. These techniques are used because removing the data from the dataset every time is not feasible and can lead to a reduction in the size of the dataset to a large extend, which not only raises concerns for biasing the dataset but also leads to incorrect analysis.

Not Sure What is Missing Data ? How it occurs? And its type? Have a look to know more about it.

Let’s understand the concept of Imputation from the above Fig {Fig 1}. In the above image, I have tried to represent the Missing data on the left table(marked in Red) and by using the Imputation techniques we have filled the missing dataset in the right table(marked in Yellow), without reducing the actual size of the dataset. If we notice here we have increased the column size, which is possible in Imputation(Adding “Missing” category imputation).

## Why Data Imputation is Important?

So, after knowing the definition of Imputation, the next question is Why should we use it, and what would happen if I don’t use it?

We use imputation because Missing data can cause the below issues:

• Incompatible with most of the Python libraries used in Machine Learning:- Yes, you read it right. While using the libraries for ML(the most common is skLearn), they donâ€™t have a provision to automatically handle these missing data and can lead to errors.
• Distortion in Dataset:- A huge amount of missing data can cause distortions in the variable distribution i.e it can increase or decrease the value of a particular category in the dataset.
• Affects the Final Model:-Â the missing data can cause a bias in the dataset and can lead to a faulty analysis by the model.

Another and the most important reason is “We want to restore the complete dataset”. This is mostly in the case when we do not want to lose any(more of) data from our dataset as all of it is important, & secondly, dataset size is not very big, and removing some part of it can have a significant impact on the final model.

Great..!! we got some basic concepts of Missing data and Imputation. Now, let’s have a look at the different techniques of Imputation and compare them. But before we jump to it, we have to know the types of data in our dataset.

Sounds strange..!!! Don’t worry… Most data is of 4 types:- Numeric, Categorical, Date-time & Mixed. These names are quite self-explanatory so not going much in-depth and describing them.

## Data Imputation Techniques

Moving on to the main highlight of this article – techniques used in data imputation.

Note:- I will be focusing only on Mixed, Numerical and Categorical Imputation here. Date-Time will be part of next article.

### Complete Case Analysis(CCA)

This is a quite straightforward method of handling the Missing Data, which directly removes the rows that have missing data i.e. we consider only those rows where we have complete data i.e. data is not missing. This method is also popularly known as “Listwise deletion”.

• Assumptions:-
• Data is Missing At Random(MAR).
• Missing data is completely removed from the table.
• Easy to implement.
• No Data manipulation required.
• Limitations:-
• Deleted data can be informative.
• Can lead to the deletion of a large part of the data.
• Can create a bias in the dataset, if a large amount of a particular type of variable is deleted from it.
• The production model will not know what to do with Missing data.
• When to Use:-
• Data is MAR(Missing At Random).
• Good for Mixed, Numerical, and Categorical data.
• Missing data is not more than 5% – 6% of the dataset.
• Data doesn’t contain much information and will not bias the dataset.

Code:

``````## To check the shape of original dataset
train_df.shape

## Output (614 rows & 13 columns)
(614, 13)

## Finding the columns that have Null Values(Missing Data)
## We are using a for loop for all the columns present in dataset with average null values greater than 0
na_variables = [var for var in train_df.columns if train_df[var].isnull().mean() > 0]

## Output of column names with null values
['Gender', 'Married', 'Dependents', 'Self_Employed', 'LoanAmount', 'Loan_Amount_Term', 'Credit_History']

## We can also see the mean Null values present in these columns {Shown in image below}
data_na = train_df[na_variables].isnull().mean()

## Implementing the CCA techniques to remove Missing Data
data_cca = train_df.dropna(axis=0)  ### axis=0 is used for specifying rows

## Verifying the final shape of the remaining dataset
data_cca.shape

## Output (480 rows & 13 columns)
(480, 13)``````

Output:

Here we can see, dataset had initially 614 rows and 13 columns, out of which 7 rows had missing data(na_variables), their mean missing rows are shown by data_na. We notice that apart from <Credit_History> & <Self_Employed> all have mean less than 5%. So as per the CCA, we dropped the rows with missing data which resulted in a dataset with only 480 rows. Around 20% of the data reduction can be seen here, which can cause many issues going ahead.

### Arbitrary Value Imputation

This is an important technique used in Imputation as it can handle both the Numerical and Categorical variables. This technique states that we group the missing values in a column and assign them to a new value that is far away from the range of that column. Mostly we use values like 99999999 or -9999999 or “Missing” or “Not defined” for numerical & categorical variables.

• Assumptions:-
• Data is not Missing At Random.
• The missing data is imputed with an arbitrary value that is not part of the dataset or Mean/Median/Mode of data.
• Easy to implement.
• We can use it in production.
• It retains the importance of “missing values” if it exists.
• Can distort original variable distribution.
• Arbitrary values can create outliers.
• Extra caution required in selecting the Arbitrary value.
• When to Use:-
• When data is not MAR(Missing At Random).
• Suitable for All.

Code:

``````## Finding the columns that have Null Values(Missing Data)
## We are using a for loop for all the columns present in dataset with average null values greater than 0

na_variables = [ var for var in train_df.columns if train_df[var].isnull().mean() > 0 ]

## Output of column names with null values
['Gender','Married','Dependents','Self_Employed','LoanAmount','Loan_Amount_Term','Credit_History']

## Use Gender column to find the unique values in the column
train_df['Gender'].unique()

## Output
array(['Male','Female',nan])

## Here nan represent Missing Data
## Using Arbitary Imputation technique, we will Impute missing Gender with "Missing"  {You can use any other value also}
arb_impute = train_df['Gender'].fillna('Missing')
arb.impute.unique()

## Output
array(['Male','Female','Missing'])``````

Output:

We can see here column Gender had 2 Unique values {‘Male’,’Female’} and few missing values {nan}. By using the Arbitrary Imputation we filled the {nan} values in this column with {missing} thus, making 3 unique values for the variable ‘Gender’.

### Frequent Category Imputation

This technique says to replace the missing value with the variable with the highest frequency or in simple words replacing the values with the Mode of that column. This technique is also referred to as Mode Imputation.

• Assumptions:-
• Data is missing at random.
• There is a high probability that the missing data looks like the majority of the data.
• Implementation is easy.
• We can obtain a complete dataset in very little time.
• We can use this technique in the production model.
• The higher the percentage of missing values, the higher will be the distortion.
• May lead to over-representation of a particular category.
• Can distort original variable distribution.
• When to Use:-
• Data is Missing at Random(MAR)
• Missing data is not more than 5% – 6% of the dataset.

Code:

``````## finding the count of unique values in Gender
train_df['Gender'].groupby(train_df['Gender']).count()

## Output (489 Male & 112 Female)
Male 489
Female 112

## Male has higgest frequency. We can also do it by checking the mode
train_df['Gender'].mode()

## Output
Male

## Using Frequent Category Imputer
frq_impute = train_df['Gender'].fillna('Male')
frq_impute.unique()

## Output
array(['Male','Female'])``````

Output:

Here we noticed “Male” was the most frequent category thus, we used it to replace the missing data. Now we are left with only 2 categories i.e. Male & Female.

Thus, we can see every technique has its Advantages and Disadvantages, and it depends upon the dataset and the situation for which different techniques we are going to use.

## Conclusion

Handling missing data is crucial for maintaining analysis integrity in data science. This article has covered key imputation methodsâ€”Complete Case Analysis (CCA), Arbitrary Value Imputation, and Frequent Category Imputationâ€”highlighting their assumptions, benefits, and drawbacks. Each method has specific applications and limitations, emphasizing the importance of context in choosing the right technique. Mastering these techniques ensures robust, accurate datasets, leading to reliable and unbiased AI models and more meaningful insights from your data.

#### Key Takeaways

• Data imputation is essential for maintaining dataset integrity and preventing biases in analysis.
• Incompatible missing data can cause issues with many machine learning libraries, leading to errors and faulty analyses.
• Complete Case Analysis (CCA) is straightforward but can result in significant data loss and potential bias.
• Arbitrary Value Imputation is versatile but requires careful selection of arbitrary values to avoid distorting the dataset.
• Frequent Category Imputation is useful for categorical data but can over-represent certain categories if not used carefully.
• Each imputation technique has its own set of advantages and limitations, making the choice context-dependent.
• Proper data imputation ensures that AI models are trained on accurate, unbiased, and comprehensive datasets.

Q1. What are the different types of single imputation?

A. The different types of single imputation include Mean Imputation, Median Imputation, Mode Imputation, and Arbitrary Value Imputation. Each method replaces missing values with a single, substituted value.

Q2. What is data imputation with mean?

A. Data imputation with mean involves replacing missing values with the mean of the available values in the dataset. This method ensures that the overall mean of the data remains unchanged.

Q3. When should you impute data?

A. Data should be imputed when missing values are present and removing these values could lead to significant data loss, potential bias, or distortion in the analysis.

Q4. What are the benefits of data imputation?

A. The benefits of data imputation include maintaining dataset integrity, reducing biases, preventing analysis distortion, and ensuring compatibility with machine learning libraries, leading to more accurate and reliable models.

The media shown in this article are not owned by Analytics Vidhya and are used at the Authorâ€™s discretion.

Shashank Singhal 10 Jun, 2024