## Introduction

There are no shortcuts for data exploration. If you are in a state of mind, that machine learning can sail you away from every data storm, trust me, it won’t. After some point of time, you’ll realize that you are strugglingÂ at improving model’s accuracy. In such situation, data exploration techniques will come to your rescue.

I can confidently say this, because I’ve been through such situations, a lot.

I have been a Business Analytics professional for close to three years now. In my initial days, one of my mentor suggested me to spend significant time on exploration and analyzing data. Following his advice has served me well.

I’ve created this tutorial to help you understand the underlying techniques of data exploration. As always, I’ve tried my best to explain these concepts in the simplest manner. For better understanding, I’ve taken up few examples to demonstrate the complicated concepts.

## Table of Contents

**Steps of Data Exploration and Preparation****Missing Value Treatment**- Why missing value treatment is required ?
- Why data has missing values?
- Which are the methods to treat missing value ?

**Techniques of Outlier Detection and Treatment**- What is an outlier?
- What are the types of outliers ?
- What are the causes of outliers ?
- What is the impact of outliers on dataset ?
- How to detect outlier ?
- How to remove outlier ?

**The Art of Feature Engineering**- What is Feature Engineering ?
- What is the process of Feature Engineering ?
- What is Variable Transformation ?
- When should we use variable transformation ?
- What are the common methods of variable transformation ?
- What is feature variable creation and its benefits ?

### Let’s get started.

## 1. Steps of Data Exploration and Preparation

Remember the quality of your inputs decide the quality of your output. So, once you have got your business hypothesis ready, it makes sense to spend lot ofÂ time and efforts here. WithÂ my personal estimate, data exploration, cleaning and preparation canÂ takeÂ up to 70% of your total project time.

Below are the steps involved to understand, clean and prepare your data for building your predictive model:

- Variable Identification
- Univariate Analysis
- Bi-variate Analysis
- Missing values treatment
- Outlier treatment
- Variable transformation
- Variable creation

Finally, we will need to iterate over steps 4 – 7 multiple times before we come up with our refined model.

Let’s now study each stage in Â detail:-

### Variable Identification

First, identify **Predictor** (Input) and **Target** (output) variables. Next, identify the data type and category of the variables.

Let’s understand this step more clearly by taking an example.

Example:- Suppose, we want to predict, whether the students will play cricket or not (refer below data set). Here you need to identify predictor variables, target variable, data type of variables and category of variables.Below, the variables have been defined in different category:

### Univariate Analysis

At this stage, we explore variables one by one. Method to perform uni-variate analysis will depend on whether the variable typeÂ is categorical orÂ continuous.Â Let’s look at these methods and statistical measures for categorical and continuous variables individually:

**Continuous Variables:-**Â In case of continuous variables, we need to understand the central tendency and spread of the variable. These are measured using various statistical metrics visualization methods as shown below:

**Note:**Â Univariate analysisÂ is also used to highlight missing and outlier values. In theÂ upcoming part of this series, we will look at methods to handle missing and outlier values. To know more about these methods, you can refer courseÂ descriptive statistics from Udacity.

**Categorical Variables:- **For categorical variables, we’ll use frequency table to understand distribution ofÂ each category. We can also read as percentage of valuesÂ under each category. It can be be measured using two metrics, **Count** and **Count%** against each category. Bar chart can be used as visualization.

### Bi-variateÂ Analysis

Bi-variate Analysis finds out the relationship between two variables. Here, we look forÂ association and disassociation between variables at a pre-defined significance level. We can perform bi-variate analysis for any combination of categorical and continuous variables. The combination can be: Categorical & Categorical, Categorical & Continuous and Continuous & Continuous. Different methods are used to tackle these combinations during analysis process.

Let’s understand the possible combinations in detail:

**Continuous & Continuous: **While doing bi-variate analysis between two continuous variables, we should look at scatter plot. It is a nifty way toÂ find out the relationship between two variables. The pattern of scatter plot indicates the relationship between variables. The relationship can be linear or non-linear.

Scatter plot shows the relationship between two variable but does not indicatesÂ the strength of relationship amongst them. To find the strength of the relationship, we useÂ Correlation. Correlation varies between -1 and +1.

- -1: perfect negative linear correlation
- +1:perfect positive linear correlation andÂ

- 0: No correlation

Correlation can be derived using following formula:

**Correlation = Covariance(X,Y) / SQRT( Var(X)* Var(Y))**

Various tools have function or functionality to identify correlation between variables. In Excel, function CORREL() is used to return the correlation between two variables and SAS uses procedure PROC CORR to identify the correlation. These function returns Pearson Correlation value to identify the relationship between two variables:

In above example, we have goodÂ positive relationship(0.65) between two variables X and Y.

**Categorical & Categorical:Â **To find the relationship between two categorical variables, we can use following methods:

**Two-way table:**We can startÂ analyzing the relationship by creating a two-way table of count and count%. The rows represents the category of one variable and the columns representÂ the categories of the other variable. We show count or count% of observations available in each combination of row and column categories.**Stacked Column Chart:**This method is more of a visual form of Two-way table.

**Chi-Square Test:**Â This test is used to derive the statistical significance of relationship between the variables. Also, it tests whether the evidence in the sample is strong enough to generalize that the relationship for a larger population as well. Chi-square is based on the difference between the expected and observed frequencies in one or more categories in the two-way table. It returns probability for the computed chi-square distribution with the degree of freedom.

Probability of 0: It indicates that both categorical variable are dependent

Probability of 1: It shows that both variables are independent.

Probability less than 0.05: It indicates that the relationship between the variables is significant at 95% confidence. The chi-square test statistic for a test of independence of two categorical variables is found by:

where *O* represents the observed frequency. *E* is the expected frequency under the null hypothesis and computed by:

From previous two-way table, the expected count for product category 1 to be of small size isÂ Â 0.22. It is derived byÂ taking the row total for SizeÂ (9) times the column total for Product category (2) then dividing by the sample size (81). This is procedure is conducted for each cell. Statistical Measures used to analyze the power of relationship are:

- Cramer’s V for Nominal Categorical Variable
- Mantel-Haenszed Chi-Square for ordinal categorical variable.

Different data science language and tools have specific methods to perform chi-square test. In SAS, we can use **Chisq**Â as an option with **Proc freq** to perform this test.

**Categorical & Continuous: **While exploring relation between categorical and continuous variables, we can draw box plots for each level of categorical variables. If levels are smallÂ in number, it will not show the statistical significance. To look at the statistical significance we can perform Z-test, T-test or ANOVA.

**Z-Test/ T-Test:-**Either test assess whether mean of two groups are statistically different from each other or not.If the probability of Z is small then the difference of two averages is more significant. The T-test is very similar to Z-test but it is used when number of observation for both categories is less than 30.

**ANOVA:-**It assesses whether the average of more than two groups is statistically different.

**Example:** Suppose, we want to test the effect of five different exercises. For this, we recruit 20 men and assign one type of exercise to 4 men (5 groups). Their weights are recorded after a few weeks.Â We need to find out whether the effect of these exercises on them is significantly different or not. This can be done by comparing the weights of the 5 groups of 4 men each.

Till here, we have understoodÂ the first three stages of Data Exploration, Variable Identification, Uni-Variate and Bi-Variate analysis. We also looked at various statistical and visual methods to identify the relationship between variables.Â

Now, we will look at the methods ofÂ Missing values Treatment.Â More importantly, we will also look at why missing values occur in our data and why treating them is necessary.

## 2. Missing Value Treatment

### Why missing values treatment is required?

Missing data in the training data set can reduce the power / fit of a model or can lead toÂ a biased model because we have not analysedÂ the behavior and relationship with other variables correctly. It can lead to wrong prediction or classification.

Notice the missing values in the image shown above: In the left scenario, we have not treated missing values. The inference from this data set is that the chances of playing cricket by males is higher than females. On the other hand, if you look at the second table, which shows data after treatment of missing values (based on gender), we can see that females have higher chances of playing cricket compared to males.

### Why my data has missing values?

We looked at the importance of treatment of missing values in a dataset. Now, let’s identify the reasons for occurrence of these missing values. TheyÂ may occur at two stages:

**Data Extraction**: It isÂ possible that there are problems with extraction process. In such cases, we should double-check for correct data with data guardians. Some hashing procedures can also be used to make sure data extraction is correct. Errors at data extraction stage are typically easy to find and can be corrected easily as well.**Data collection**: These errors occurÂ at time of data collection and are harder to correct. They can be categorized in four types:**Missing completely at random:**Â This is a case when the probability of missing variable is same for all observations. For example: respondents of data collection process decide that they will declare their earning after tossing a fair coin. If an head occurs, respondentÂ declares his / her earnings & vice versa. Here each observation has equal chance of missing value.**Missing at random:**This is a case when variable is missing at random and missing ratio varies for different values / level of other input variables. For example:Â We are collecting data for age and female has higher missing value compare to male.**Missing that depends on unobserved predictors:**Â This is a case when the missingÂ values are not random and areÂ related toÂ the unobserved input variable. For example: In a medical study, if a particular diagnostic causes discomfort, then there is higher chance of drop out from the study. This missing value is not at random unless we have included â€śdiscomfortâ€ť as an input variable for all patients.**Missing that depends on the missing value itself:**This is a case when the probability of missing value is directly correlated with missing value itself. For example: People with higher or lower income are likely to provide non-response to their earning.

### Which are the methods to treat missing values ?

**Deletion:Â**It is of two types: List Wise Deletion and Pair Wise Deletion.- In list wise deletion, we delete observations where any of the variable is missing. Simplicity is one of the major advantage of this method, but this method reduces the power of model because it reduces the sample size.
- In pair wise deletion, we perform analysisÂ with all cases in which the variables of interest are present. Advantage of this method is, it keeps as many cases available for analysis. One of the disadvantage of this method, it uses different sample size for different variables.

- Deletion methods are used when the nature of missing data is “
**Missing completely at random**” else non random missing values can bias the model output.

**Mean/ Mode/ Median Imputation:Â**Imputation is a methodÂ to fill in the missing values with estimated ones. The objective is to employ known relationships that can be identified in the valid values of the data set to assist in estimating the missing values. Mean / Mode / Median imputation is one of the most frequently used methods. It consists of replacing the missing data for a given attribute by the mean or median (quantitative attribute) or mode (qualitative attribute) of all known values of that variable. It can be of two types:-**Generalized Imputation:**In this case, we calculate the mean or median for all non missing values of that variable then replace missing value with mean or median. Like in above table, variable “**Manpower”**is missing so we take average of all non missing values of “**Manpower”**Â (**28.33**) and then replace missing value with it.**Similar case Imputation:**In this case, we calculate average for gender “**Male”Â**(29.75) and “**Female**” (25) individually of non missing values then replace the missing value based on gender. For “**Male**“, we will replace missing values of manpower with 29.75 and for “**Female**” with 25.

**Prediction Model**: Â Prediction model is one of the sophisticated methodÂ for handling missing data. Here, weÂ create a predictive model to estimate values that will substitute the missing data. Â In this case, we divide our data set into two sets: One set with no missing values for the variable and another one with missing values. First data set become training data set of the model while second data set with missing values is test data set and variable with missing values is treated as target variable. Next, we create a model to predict target variable based on other attributes of the training data set and populate missing values of test data set.We can use regression, ANOVA, Logistic regression and various modeling technique to perform this. There are 2 drawbacks for this approach:- The model estimated values are usually more well-behaved than the true values
- If there are no relationships withÂ attributes in the data set and the attribute with missing values, then the model will not be precise for estimating missing values.

**KNN Imputation:**In this method of imputation, the missing values of anÂ attributeÂ are imputed using the given number of attributes that are most similar to the attribute whose values are missing. The similarity of two attributes is determined using a distance function. It is also known to have certain advantage & disadvantages.**Advantages:**- k-nearest neighbour can predict both qualitative & quantitative attributes
- Creation of predictive model for each attribute with missing data is not required
- Attributes with multiple missing values can be easily treated
- Correlation structure of the data is taken into consideration

**Disadvantage:**- KNN algorithm is very time-consuming in analyzing large database. It searches through all the dataset looking for the most similar instances.
- Choice of k-value is very critical. Higher value of k would include attributes which are significantly different from what we need whereas lower value of k implies missing out of significant attributes.

After dealing with missing values, the next task is to deal with outliers. Often, we tend to neglect outliers while building models. This is a discouraging practice. Outliers tend to make your data skewed and reduces accuracy. Let’s learn more about outlier treatment.

## 3. Techniques of Outlier Detection and Treatment

### What is an Outlier?

Outlier is a commonly usedÂ terminology by analysts and data scientists asÂ it needsÂ close attention else itÂ can result in wildly wrongÂ estimations. Simply speaking,Â Outlier is an observation that appears far away and diverges from an overall pattern in a sample.

Let’s take an example, we do customer profiling and find out that the average annual income of customers is $0.8 million. But, there are two customers having annual income of $4 and $4.2 million. These two customers annual income is much higher thanÂ rest of the population. These two observations will be seen asÂ Outliers.

### What are the types of Outliers?

Outlier can be of two types:Â **Univariate** andÂ **Multivariate**. Above, we have discussed the example of univariate outlier. These outliers can be found when we look at distribution of a single variable. Multi-variate outliers are outliers in an n-dimensional space. In order to find them, you have to look at distributions in multi-dimensions.

Let us understand this with an example. Let us say we are understanding the relationship between height and weight. Below, we have univariate and bivariate distribution for Height, Weight. Take a look atÂ the box plot. We do not have any outlier (above and below 1.5*IQR, most common method). Now look at the scatter plot. Here, we have two valuesÂ belowÂ and one above the average in a specific segment of weight and height.

### What causes Outliers?

Whenever we come across outliers, the ideal way to tackle them is to find out the reason of having these outliers. The method to deal with them would then depend on the reason of their occurrence. Causes of outliers can be classified in two broad categories:

**Artificial (Error) / Non-natural****Natural**.

Let’s understand various types of outliers in more detail:

**Data Entry Errors:-**Â Human errors such as errors caused during data collection, recording, or entry can cause outliers in data. For example: Annual income of a customer is $100,000. Accidentally, the data entry operator puts an additional zero in the figure. Now the income becomesÂ $1,000,000 which is 10 times higher. Evidently, this will be theÂ outlier value when compared with rest of the population.**Measurement Error:Â**It is the most common source of outliers. This is caused when the measurement instrument used turns out to be faulty. For example: There are 10 weighing machines. 9 of them are correct, 1 is faulty. Weight measured by people on the faulty machine will be higher / lower than the rest of people in the group.Â The weights measured on faulty machine can lead to outliers.**Experimental Error:**Another cause of outliers is experimental error. For example: In a 100m sprint of 7 runners, one runner missed out on concentrating on the ‘Go’ callÂ which caused him to start late. Hence, this causedÂ the runner’s run time to beÂ moreÂ than other runners. His total run time canÂ be an outlier.**Intentional Outlier**This is commonly found in self-reported measures that involves sensitive data. For example: Teens would typically under report the amount of alcohol that they consume. Only a fraction of them would report actual value. Here actual values might look like outliers because rest of the teens are under reporting the consumption.**:Â****Data Processing Error:**Whenever we perform data mining, we extract data from multiple sources. It is possible that some manipulation or extraction errors may lead to outliers in the dataset.**Sampling error:**Â For instance, we have to measure the height of athletes. By mistake, we include a few basketball players in the sample. This inclusion is likely to cause outliers in the dataset.**Natural Outlier:Â**When an outlier is not artificial (due to error), it isÂ a natural outlier. For instance: In my last assignment with one of the renowned insurance company, I noticed that the performance of top 50 financial advisors was far higher than rest of the population. Surprisingly, it was not due to any error. Hence, whenever we perform any data mining activity with advisors, we used to treat this segment separately.

### What is the impact of Outliers on a dataset?

Outliers can drastically change the results of the data analysis and statistical modeling. There are numerous unfavourable impacts of outliers in the data set:

- ItÂ increases the error variance and reduces the power of statistical tests
- If the outliers are non-randomly distributed, they can decrease normality
- TheyÂ can bias or influence estimates that may be of substantive interest
- TheyÂ can also impact the basic assumption of Regression, ANOVA and other statistical model assumptions.

To understand the impact deeply, let’s take an example to checkÂ what happens to a data set with and without outliersÂ in theÂ data set.

**Example:**

As you can see, data set with outliers has significantly different mean and standard deviation. In the first scenario, we will say that average is 5.45. But with the outlier, average soarsÂ toÂ 30. This would change the estimate completely.

### How to detect Outliers?

Most commonly used method to detect outliers is visualization. We use various visualization methods, like **Box-plot**, **Histogram**, **Scatter Plot** (above, we have used box plot and scatter plot for visualization). Some analysts also various thumb rules to detectÂ outliers. Some of them are:

- Any value, which is beyond the range of -1.5 x IQR to 1.5 x IQR
- Use capping methods. Any value which out of range of 5th and 95th percentile can be considered as outlier
- Data points, three or more standard deviation away from mean are considered outlier
- Outlier detection is merely a special case of the examination of data for influential data points and it also depends on the business understanding
- Bivariate and multivariate outliers are typically measured using either an index of influence or leverage, or distance. Popular indices such as Mahalanobisâ€™ distance and Cookâ€™s
*D*are frequently used to detect outliers. - In SAS, we can use PROC Univariate, PROC SGPLOT. To identify outliers and influential observation, we also look at statistical measure like STUDENT, COOKD, RSTUDENT and others.

### How to remove Outliers?

Most of the waysÂ to dealÂ with outliers are similar to the methods ofÂ missing values like deleting observations, transforming them, binning them, treat them as a separate group, imputing values and other statistical methods. Here, we willÂ discuss theÂ common techniques used to deal with outliers:

**Deleting observations:Â **We delete outlier values if it is due to data entry error, data processing error or outlier observations are very small in numbers. We can also use trimming at both ends to remove outliers.

**Transforming and binning values:Â **Transforming variables can also eliminate outliers. Natural log of a value reduces the variation caused by extreme values. Binning is also a form of variable transformation. Decision Tree algorithm allowsÂ to deal with outliers well due to binning of variable. We can also use the process of assigning weights to different observations.

**Imputing:Â **LikeÂ imputationÂ of missing values, we can also imputeÂ outliers. We can use mean, median, mode imputation methods. Before imputing values, we should analyse if it is natural outlier or artificial. If it is artificial, we can go with imputing values. We can also use statistical model to predict values of outlier observation and after that we can impute it with predicted values.

**Treat separately:Â **If there are significant number of outliers, we should treat them separatelyÂ in the statistical model. One of the approach is to treat both groups as two different groups and build individual model for both groups and thenÂ combine the output.

Till here, we have learnt about steps of data exploration, missing value treatment and techniques of outlier detection and treatment. These 3 stages will make your raw data better in terms of information availability and accuracy. Let’s now proceed to the final stage of data exploration. It is Feature Engineering.

## 4. The Art of Feature Engineering

### What is Feature Engineering?

Feature engineering is the science (and art) of extracting more information from existing data. You are not adding any new data here, but you are actually making the data you already have more useful.

For example, let’s say you are trying to predict foot fall in a shopping mall based on dates. If you try and use the dates directly, you may not be able to extract meaningful insights from the data. This is because the foot fall is less affected by the day of the month than it is by the day of the week. Now this information about day of week is implicit in your data. You need to bring it out to make your model better.

This exercising of bringing out information from data in known as feature engineering.

### What is the process of Feature Engineering ?

You perform feature engineering once you have completed the first 5 steps in data exploration –Â Variable Identification,Â Univariate, Bivariate Analysis,Â Missing Values ImputationÂ andÂ Outliers Treatment. Feature engineering itself can be divided in 2 steps:

- Variable transformation.
- Variable / Feature creation.

These two techniques are vital in data exploration andÂ haveÂ a remarkableÂ impact on the power of prediction.Â Let’sÂ understand each of this step in more details.

### What is Variable Transformation?

In data modelling, transformation refers toÂ the replacement of a variable by a function. For instance, replacing a variable x by the square / cube root or logarithm x is a transformation. In other words, transformation is a process that changes the distribution or relationship of a variable with others.

Letâ€™s look at the situations when Â variable transformation is useful.

### When should we use Variable Transformation?

Below are the situations where variable transformation is aÂ requisite:

- When we want to
**change the scale**of a variable or standardize the values of a variable for better understanding. While this transformation is a must if you have data in different scales, this transformation does not change the shape of the variable distribution

- When we can
**transform complex non-linear relationships into linear relationships**. Existence of a linear relationship between variables isÂ easier to comprehend compared to a non-linear or curved relation. Transformation helps us to convert a non-linear relation into linear relation. Scatter plot can be used to find the relationship between two continuous variables. These transformations also improve theÂ prediction. Log transformation is one of the commonly used transformation technique used in these situations.

**Symmetric distribution is preferredÂ over skewed distribution**as it is easier to interpret and generate inferences. Some modeling techniquesÂ requires normal distribution of variables. So, whenever we have a skewed distribution, we can useÂ transformations whichÂ reduce skewness. For right skewed distribution, we take square / cube root or logarithm of variable and for left skewed, we take square / cube or exponential of variables.

- Variable Transformation is also done from an
**implementationÂ point of view**(Human involvement). Let’s understand it more clearly. In one of my project on employee performance, I found that age has direct correlation with performance of the employeeÂ i.e. higher the age, betterÂ the performance. From an implementation stand point, launching age based progamme might present implementation challenge. However, categorizing the sales agents in three age group buckets of <30 years, 30-45 years and >45 Â and then formulatingÂ three different strategies for each group is a judicious approach. This categorization technique is known as Binning of Variables.

### What are the common methods of Variable Transformation?

There are various methods used to transform variables. As discussed, some of them include square root, cube root, logarithmic, binning, reciprocal and many others. Letâ€™s look at these methods in detailÂ byÂ highlightingÂ the pros and cons of these transformation methods.

**Logarithm:**Log of a variable is aÂ common transformation method used to change the shape of distribution of the variable on a distribution plot. It is generallyÂ used for reducing right skewness of variables. Though, It canâ€™t be applied to zero or negative values as well.

**Square / Cube root:**The square and cube root of a variable has a soundÂ effect on variable distribution. However, it is not as significant as logarithmic transformation. Cube root has its own advantage.Â It can be applied to negative values includingÂ zero. Square root can be applied to positive values including zero.

**Binning:**It is used to categorize variables. It is performed on original values, percentile or frequency. Decision of categorization technique is based on business understanding. For example, we can categorize income in three categories, namely: High, Average and Low.**Â**We can also perform co-variate binning whichÂ depends on the value of more than one variables.

### What is Feature /Â Variable Creation & its Benefits?

Feature / Variable creation is a process to generate a new variables / features based on existing variable(s). For example, say, weÂ have date(dd-mm-yy) as an input variable in aÂ data set. We can generate new variables like day, month, year, week, weekday that may have better relationship with target variable. This step is used to highlight the hidden relationship in a variable:

There are various techniques to create new features. Letâ€™s look at the some of the commonly used methods:

**Creating derived variables:**This refers toÂ creating new variables from existing variable(s) using set of functions or different methods. Letâ€™s look at it through â€ś**Titanic – Kaggle competition**â€ť. In this data set, variable age has missing values. To predict missing values, weÂ usedÂ the salutation (Master, Mr, Miss, Mrs) of name as a new variable. How do we decide which variable to create? Honestly, this depends on business understanding of the analyst, his curiosity and the set of hypothesis he might have about the problem.Â Methods such as taking log of variables, binning variables and other methods of variable transformation can also be used to create new variables.

**Creating dummy variables:**One of the most common application of dummy variable isÂ toÂ convert categorical variable into numerical variables. Dummy variables are also called Indicator Variables. It is useful to take categorical variable as a predictor in statistical models.Â Categorical variable can take values 0 and 1. Let’s take a variable ‘gender’. We can produce two variables, namely, â€ś**Var_Male**â€ť with values 1 (Male) and 0 (No male) and â€ś**Var_Female**â€ť with values 1 (Female) and 0 (No Female). We can also create dummy variables for more than two classes of a categorical variables with n or n-1 dummy variables.

#### For further read, here is a list of transformation / creation ideas which can be applied to your data.

## End Notes

As mentioned in the beginning, quality and efforts invested in data exploration differentiates a good model from a bad model.

This ends our guideÂ on data exploration and preparation. In this comprehensive guide, we looked at the seven steps of data exploration in detail. The aim of this series was to provide an in depth and step by step guide to an extremely important process in data science.

Personally, I enjoyed writing this guideÂ and would love to learn from your feedback. DidÂ you findÂ this guideÂ useful? I would appreciate yourÂ suggestions/feedback.Â Please feel free to ask your questions through comments below.

Really useful and comprehensive, thanks

Hi Ray,

I would like to thank you very much for this useful post

I took more than 30 statistical courses but your post has summarized them for me

Now all things are clear about EDA

I’m member of the John Hopkins University Data Scientists (Coursera) Group

Best,

Excellent series of blog posts. Thanks and keep up the good work!

Superb writing, crisp and comprehensive. Certainly a good refresher. Keep writing!

Very comprehensive. Thanks

Excellent article on the most important aspects of Machine Learning. The points are explained in a simple and concise manner. Thank you.

Thank you very much for this tutorial!

I haven’t come across any other article as detailed as this one. Anyone who is keen about data exploration and Predictive Analytics in general has to go through this. Wondering if you have any data set where in I can work on it.

Bookmarked!

Hi Ray,

This is a great post. You have treated a fairly vast topic with just the right amount of detail. This makes it very useful, and also very intresting. Thank you for the good work. Keep it up.

Thank you so much for this very valuable post. I like your blogs, Please continue your good work !

I would like to thank Mr. Sunil Ray for such comprehensive information. Also, I would request some to write a blog on ETL, SAS BI and how SAS BI is better than other BI tools like Tableau, Qlikview….gaining more popularity in market.

Thanks again for sharing helpful information!!

Well defined process of data exploration Sunil. I appreciated if you continue this wonderful work and post an example of data analysis step by step using Python.

Thanks

Thank you Mr. Ray for the very comprehensive discussion on data exploration. I specially liked how you emphasized on the importance of EDA with this statement “quality and efforts invested in data exploration differentiates a good model from a bad model”. Great work Sir! I wish you can tackle dimensionality reduction techniques, principal components analysis, discriminant analysis and the likes in the future. Thanks again Mr. Ray.

I found myself nodding my noggin all the way thguorh.

Its really worth to read. Very comprehensive and easy to understand . I will be happy to read your article using R on data exploration & Data preparation.

Thank you all for exciting comments and Iâ€™m glad it helped.

Regards,

Sunil

Hi Sunil: Thanks for your article on such an important topic. BTW, there is a missing graph on the paragraph Continuous & Continuous under Bi-Variate Analysis. Could you please edit it and add the missing graph. I think is pointing to a wrong place looking for the ping file. Thanks.

The missing picture/draw might be located here:

http://www.analyticsvidhya.com/wp-content/uploads/2015/02/Data_exploration_4.png

This picture is the missing one below the paragraph:

“Continuous & Continuous: While doing bi-variate analysis between two continuous variables, we should look at scatter plot. It is a nifty way to find out the relationship between two variables. The pattern of scatter plot indicates the relationship between variables. The relationship can be linear or non-linear.”

Hi Sunil,

An intriguing article, I can see the amount of hard work you must have put into it. Its a must read.

Thanks,

Akshay Kher

Thanks a lot for the comprehensive material Sunil. I had All these points scattered across but you got all of them together, along with few new pointers. Bookmarked this page and this would now be my first page to refer for any data analysis project.

Clear explanation with example and graph.

Thanks.

This common techniques are core of any data analytics project. Good work keep up.

Very well explained and interesting article..It helped me a lot….Thanks a lot

when we create new variable like var_male and var_female we assign 0,1 to them? how is this 0,1 is used in our model? can we assign 200 instead of 0 and 2000 instead of 1?

Please help .

clear, Concise and Very well explained. !!

Great article.

One quick suggestion regarding log transform for zero or negative values.

For all values, convert to absolute value, add one to all values (if data has lots if zeros), take log, then finally reapply the negative sign where original was negative.

E.g. log(-2) = -1Ă—(log(abs(-2)+1))

Hope that helps.

Excellent guide! Thank you very much! Very pedagogic and comprehensive. Two thumbs up!

An excellent place to come back when starting a new data project…

Very well explained article.. Person having basic math /statistics understanding can also understand subject well..

Concise and comprehensive. Great article.

Very well written.

One of the best blogs I have ever read till date!

This is really great. Thank you so much!!

Can we use Weight of Evidence to impute outliers and Missing Values??

Great! Thanks!

Very Useful. Thank you.. đź™‚

Amazing guide.. very structured and simplistic. enjoyed and learnt a lot reading this article.

Many thanks for the guide, very useful. Would you advise R packages that help with data exploration?

Thanks

THANK YOU FOR SHARING THIS CONCEPTS AND METHOD.

If a variable is very skewed at 0 but valid. How should we treat them in a logistic regression framework?

Very useful.

precise and clear

Thank you.

Excellent article

Thank you very much

really awesome..crisp and concise

I open a file in google drive to keep this page alone as a cheatsheet…Thank you so much..

This is very useful summary, thank you for that!

I particularly liked the before-after comparisons to demonstrate the importance of the process steps.

Thanks,

Chill

Great article! Few questions:

1) Do you run your data exploration on sample or full data set? If sample then what percentage and any article on how to take samples for unstructured text based dataset.

2) How to explore fields which are unstructured text, images etc. Do we need to run feature extraction before we explore. how do we explore them anyway?

I understand there’s no single answer but in your opinion what’s the best way to explore unstructured dataset.

The blog articles from AV are just awesome! Thanks to all the blog writers for sharing their knowledge.

Definitely going to Bookmark this blog ! Thank you .

you nailed the process. I thoroughly enjoyed reading your blog and learned a lot!!!! Thanks a lot for investing time and sharing your experience.

Well Written. it really shows how to tackle the data

Excellent read on EDA simple and to the point. Great Help to newbie like me.

Thank you for sharing knowledge. It helps a lot.

great article! Very useful!

You Sir are amazing…

Great article!

I would like to add or comment on the imputation of missing values. I once had a dataset with missing values in one of the categorical variables. Instead of replacing missing values with the most frequent value of that variable, I looked at the distribution of unique values and found that they were all uniformly distributed. With this information, I would replace a missing value by randomly choosing a value among the set of unique values. It worked quite well but I would love to hear if this was statistically the right thing to do?

The Best. Period.

Hi Sunil

Thank you very much for really useful and clear structure.

Great explanation, would be better. If you could give us some sample data and then explain step by step on that.

Loved reading it.

Thanks for sum it up in the best explanatory manner. đź™‚

Best,

Well summarised explanations covering each topic of data exploration with enough details to understand. Thanks a lot for this post.

This is such an amazing resource. Thank you very much for sharing

It was a crisp and clear and more importantly step by step explanation of EDA process.

I read all these things here and there but first time as an organized flow.

Keep up the good work sir.You understood the pain points of novice data scientist.

great one.

Could you please also add python sample code for these examples? Thank you.

Best guide ever!