Welcome to the world of data science! Data science is the art of using data to find answers and draw meaningful insights. But how do we do that reliably? The secret sauce is statistics. It’s the powerful framework that allows us to move beyond simple observations to make confident, data-driven decisions. Whether you’re building a machine learning model or presenting findings to your team, a strong statistical foundation is non-negotiable. In this article, we will learn all the important statistical concepts that are required for Data Science roles.
Imagine you’re cooking a large pot of soup. The entire pot is your population. It’s the complete group you’re interested in, whether it’s all your app users or every home in a city. Since you can’t possibly test the whole pot, you take a spoonful to taste. That spoonful is your sample: a smaller, manageable subset of the population that you collect data from. The core idea is that by carefully analyzing the sample, we can make an educated guess about the entire population without having to study every single member.
A parameter is a number that describes the data from the population. And a statistic is a number that describes the data a Continuing our soup analogy, a parameter is the true value that describes the entire population. It’s the exact, real average saltiness of the whole pot of soup, a value we might never know perfectly. In data science, this could be the true average spending of all customers. A statistic, on the other hand, is a number that describes your sample. It’s the saltiness you measured in your spoonful. We use this statistic to make an educated guess about the population’s parameter.
Statistics is a discipline that concerns the collection, organization, analysis, interpretation, and presentation of data.
It means, as part of statistical analysis, we collect, organize, and draw meaningful insights from the data either through visualizations or mathematical explanations.
Statistics is broadly categorized into two types:
As the name suggests, in Descriptive statistics, we describe the data using the Mean, Standard deviation, Charts, or Probability distributions.
Basically, as part of descriptive Statistics, we measure the following:
In Inferential statistics, we estimate the population parameters. Or we run Hypothesis testing to assess the assumptions made about the population parameters.
In simple terms, we interpret the meaning of the descriptive statistics by applying them to the population.
For example, we are conducting a survey on the number of two-wheelers in a city. Assume the city has a total population of 5L people. So, we take a sample of 1000 people, as it is impossible to run an analysis on the entire population data.
From the survey conducted, it is found that 800 people out of 1000 (800 out of 1000 is 80%) are two-wheelers. So, we can infer these results to the population and conclude that 4L people out of the 5L population are two-wheelers.
At a higher level, data is categorized into two types: Qualitative and Quantitative.
Qualitative data is non-numerical. Some of the examples are eye colour, car brand, city, etc.
On the other hand, Quantitative data is numerical, and it is again divided into Continuous and Discrete data.
Discrete data is again divided into Categorical and Count Data.

In statistics, the level of measurement is a classification that describes the relationship between the values of a variable.
We have four fundamental levels of measurement. They are:
1. Nominal Scale: This scale contains the least information since the data have names/labels only. It can be used for classification. We cannot perform mathematical operations on nominal data because there is no numerical value to the options (numbers associated with the names can only be used as tags).
Example: Which country do you belong to? India, Japan, Korea.
2. Ordinal Scale: In comparison to the nominal scale, the ordinal scale has more information because, along with the labels, it has order/direction.
Example: Income level – High income, medium income, low income.
3. Interval Scale: It is a numerical scale. The Interval scale has more information than the nominal, ordinal scales. Along with the order, we know the difference between the two variables (interval indicates the distance between two entities).
Mean, median, and mode can be used to describe the data.
Example: Temperature, income, etc.
4. Ratio Scale: The ratio scale has the most information about the data. Unlike the other three scales, the ratio scale can accommodate a true zero point. The ratio scale is simply said to be the combination of Nominal, Ordinal, and Interval scales.
Example: Current weight, height, etc.
We have four moments of business decisions that help us understand the data.
(It is also known as the First Moment of Business Decision)
Talks about the centrality of the data. To keep it simple, it is a part of descriptive statistical analysis where a single value at the centre represents the entire dataset.
The central tendency of a dataset can be measured using:
(It is also known as the Second Moment Business Decision)
Talks about the spread of data from its centre.
Dispersion can be measured using:
Measure |
Population |
Sample |
|---|---|---|
| Mean | µ = (Σ Xi)/N | x̄ = (Σ xi)/n |
| Median | The middle value of the data | The middle value of the data |
| Mode | Most occurred value | Most occurred value |
| Variance | σ2 = (Σ Xi – µ)2/N | s2 = (Σ xi – x̄)2/ (n-1) |
| Standard Deviation | σ = sqrt((Σ Xi – µ)2/N) | s = sqrt((Σ xi – x̄)2/ (n-1)) |
| Range | Max-Min | Max-Min |
(It is also known as the Third Moment Business Decision)
It measures the asymmetry in the data. The two types of Skewness are:
The formula of Skewness is E [(X – µ)/ σ ]) 3 = Z3


(It is also known as Fourth Moment Business Decision)
Talks about the central peakedness or fatness of tails. The three types of Kurtosis are:
The formula of Kurtosis is E [(X – µ)/ σ ]) 4-3 = Z4– 3

Together, Skewness and Kurtosis are called Shape Statistics.
Instead of analyzing the entire population data, we always take out a sample for analysis. The problem with sampling is that “sample means are a random variable – they vary for different samples”. And the random sample we draw can never be an exact representation of the population. This phenomenon is called sample variation.
To nullify the sample variation, we use the central limit theorem. And according to the Central Limit Theorem:
In statistical terms, a distribution is a function that shows the possible values for a variable and how often they occur. Understanding the most common distributions is like having a toolkit for describing different types of real-world phenomena.
While there are many distributions, a data scientist should be very familiar with these three:
To learn more about probability distributions, read this article.
Graphical representation refers to the use of charts or graphs to visualize, analyze, and interpret numerical data.
For a single variable (Univariate analysis), we have a bar plot, line plot, frequency plot, dot plot, boxplot, and the Normal Q-Q plot.
We will be discussing the Boxplot and the Normal Q-Q plot.
A boxplot is a way of visualizing the distribution of data based on a five-number summary. It is used to identify the outliers in the data.
The five numbers are minimum, first Quartile (Q1), median (Q2), third Quartile (Q3), and maximum.
The box region will contain 50% of the data. The lower 25% of the data region is called the Lower Whisker, and the upper 25% of the data region is called the Upper Whisker.
The Interquartile range (IQR) is the difference between the third and first quartiles. IQR = Q3 – Q1.
Outliers are the data points that lie below the lower whisker and beyond the upper whisker.
The formula to find the outliers is Outlier = Q ± 1.5*(IQR)
The outliers that lie below the lower whisker are given as Q1 – 1.5 * (IQR)
The outliers that lie beyond the upper whisker are given as Q3 + 1.5 * (IQR)

Check out my article on detecting outliers using a boxplot.
A Normal Q-Q plot is a kind of scatter plot that is plotted by creating two sets of quantiles. It is used to check if the data follows normality or not.

On the x-axis, we have the Z-scores, and on the y-axis, we have the actual sample quantiles. If the scatter plot forms a straight line, the data is said to be normal.
Hypothesis testing in statistics is a formal way to test an assumption we have made about a population. Think of it as the scientific method for data: you have a theory, and you need to see if the data supports it. The process feels a lot like a courtroom trial.
The Core Idea: In a trial, the defendant is “innocent until proven guilty.” In hypothesis testing, we have a “status quo” idea that we assume is true until our data provides enough evidence to convince us otherwise.
Here are the key players:
To learn more about hypothesis testing, read this article.
When we calculate a statistic from a sample (like the sample mean), we know it’s just an estimate of the true population parameter. If we took a different sample, we’d get a slightly different mean. A confidence interval gives us a way to deal with this uncertainty.
Instead of a single-number estimate (a “point estimate”), a confidence interval provides a range of plausible values for the population parameter.
For example, instead of saying, “Based on our sample, the average customer satisfaction score is 8.2,” we can say:
“We are 95% confident that the true average customer satisfaction score for all customers is between 7.9 and 8.5.”
This is much more powerful. The 95% confidence level means that if we were to repeat our sampling process 100 times and create 100 intervals, we would expect about 95 of those intervals to contain the true population mean. It tells us how reliable our estimation method is.
So far, we have mostly discussed analyzing a single variable (univariate analysis). However, much of data science is about understanding the relationship between variables. The simplest measure of this is correlation.
Correlation measures the strength and direction of a linear relationship between two quantitative variables. The result is a correlation coefficient, a number that is always between -1 and +1.
Think of these concepts not as separate rules to memorize, but as a connected toolkit. You now have the tools not just to describe what your data looks like (with means and standard deviations), but to ask it meaningful questions (with hypothesis tests) and understand the relationships hiding inside (with correlation).
This isn’t just about crunching numbers; it’s about building intuition and telling a compelling story with data. The journey doesn’t end here. This foundation is your launchpad into the exciting world of machine learning.
This article was published as part of the Data Science Blogathon
The media shown in this article are not owned by Analytics Vidhya and are used at the Author’s discretion.