GATE DA 2026: Study Material and Practice Questions

Vasu Deo Sankrityayan Last Updated : 16 Jan, 2026
8 min read

The yearly GATE exam is right around the corner. For some this was a long time coming—for others, a last minute priority. Whichever group you belong to, preparation would be the only focus for you now. 

This article is here to assist with those efforts. A curated list of GATE DA learning material that would get you the right topics required for overcoming the exam. 

The learning is supplemented with questions that put to test your standing and proficiency in the exam.

GATE DA: Decoded

GATE DA is the Data Science and Artificial Intelligence paper in the GATE exam that tests mathematics, programming, data science, machine learning, and AI fundamentals. Here’s the syllabus for the paper:

GATE DA Syllabus: https://gate2026.iitg.ac.in/doc/GATE2026_Syllabus/DA_2026_Syllabus.pdf

To summarize, the paper consists of the following subjects:

  1. Probability and Statistics
  2. Linear Algebra
  3. Calculus and Optimization
  4. Machine Learning
  5. Artificial Intelligence

If you’re looking for resources on a specific subject, just click on one of the above links to get to the required section.  

1. Probability and Statistics

YAY Statistics
Source: Giphy

Probability and Statistics builds the foundation for reasoning under uncertainty, helping you model randomness, analyze data, and draw reliable inferences from samples using probability laws and statistical tests.

Articles:

  • Statistics and Probability: This sets the mental model. What is randomness? What does a sample represent? Why do averages stabilize? Read this to orient yourself before touching equations.
  • Basics of Probability: This is where intuition meets rules. Conditional probability, independence, and Bayes are introduced in a way that mirrors how they appear in exam questions.
  • Introduction to Probability Distributions: Once probabilities make sense, distributions explain how data behaves at scale.

Video learning: If you prefer a guided walkthrough or want to reinforce concepts visually, use the following YouTube playlist: Probability and Statistics

Questions (click to expand)

Q1. Two events A and B are independent. Which statement is always true?

P(A ∩ B) = P(A) + P(B) P(A ∩ B) = P(A)P(B)
P(A | B) = P(B | A) P(A ∪ B) = 1
Click here to view the answer

Correct option: P(A ∩ B) = P(A)P(B)

Independence means the joint probability equals the product of marginals.

Q2. Which distribution is best suited for modeling the number of arrivals per unit time?

Binomial Poisson
Normal Uniform
Click here to view the answer

Correct option: Poisson

Poisson models counts of independent events in a fixed interval (time/space).

Q3. If X and Y are uncorrelated, then:

X and Y are independent Cov(X, Y) = 0
Var(X + Y) = Var(X) − Var(Y) E[X|Y] = E[X]
Click here to view the answer

Correct option: Cov(X, Y) = 0

Uncorrelated means covariance is zero. Independence is stronger and doesn’t automatically follow.

Q4. Which theorem explains why sample means tend to be normally distributed?

Bayes Theorem Central Limit Theorem
Law of Total Probability Markov Inequality
Click here to view the answer

Correct option: Central Limit Theorem

The CLT says the distribution of sample means approaches normal as sample size increases (under broad conditions).

If you can reason about uncertainty and variability, the next step is learning how data and models are represented mathematically, which is where linear algebra comes in.

2. Linear Algebra

Linear Algebra
Source: Giphy

Linear Algebra provides the mathematical language for data representation and transformation, forming the core of machine learning models through vectors, matrices, and decompositions.

Articles:

Video learning: If visual intuition helps, use the following YouTube playlist to see geometric interpretations of vectors, projections, and decompositions in action: Linear Algebra

Questions (click to expand)

Q1. If a matrix A is idempotent, then:

A² = 0 A² = A
Aᵀ = A det(A) = 1
Click here to view the answer

Correct option: A² = A

Idempotent matrices satisfy A² = A by definition.

Q2. Rank of a matrix equals:

Number of rows Number of linearly independent rows
Determinant Trace
Click here to view the answer

Correct option: Number of linearly independent rows

Rank is the dimension of the row (or column) space.

Q3. SVD of a matrix A decomposes it into:

A = LU A = UΣVᵀ
A = QR A = LDLᵀ
Click here to view the answer

Correct option: A = UΣVᵀ

SVD factorizes A into orthogonal matrices U, V and a diagonal matrix Σ of singular values.

Q4. Eigenvalues of a projection matrix are:

Any real numbers Only 0 or 1
Only positive Only negative
Click here to view the answer

Correct option: Only 0 or 1

Projection matrices are idempotent (P² = P), which forces eigenvalues to be 0 or 1.

With vectors and matrices in place, the focus shifts to how models actually learn by adjusting these quantities, a process governed by calculus and optimization.

3. Calculus and Optimization

Calculus and Optimization
Source: Giphy

This section explains how models learn by optimizing objective functions, using derivatives and gradients to find minima and maxima that drive training and parameter updates.

Articles:

  • Mathematics Behind Machine Learning: This builds intuition around derivatives, gradients, and curvature. It helps you understand what a minimum actually represents in the context of learning.
  • Mathematics for Data Science: This connects calculus to algorithms. Gradient descent, convergence behavior, and second-order conditions are introduced in a way that aligns with how they appear in exam and model-training scenarios.
  • Optimization Essentials: Optimization is how models improve. The essentials of optimization, from objective functions to iterative methods, and shows how these ideas drive learning in machine learning systems.

Video learning: For step-by-step visual explanations of gradients, loss surfaces, and optimization dynamics, refer to the following YouTube playlist: Calculus and Optimization

Questions (click to expand)

Q1. A necessary condition for f(x) to have a local minimum at x = a is:

f(a) = 0 f′(a) = 0
f″(a) < 0 f′(a) ≠ 0
Click here to view the answer

Correct option: f′(a) = 0

A local minimum must occur at a critical point where the first derivative is zero.

Q2. Taylor series is primarily used for:

Solving integrals Function approximation
Matrix inversion Probability estimation
Click here to view the answer

Correct option: Function approximation

Taylor series approximates a function locally using its derivatives at a point.

Q3. Gradient descent updates parameters in which direction?

Along the gradient Opposite to the gradient
Random direction Orthogonal direction
Click here to view the answer

Correct option: Opposite to the gradient

The negative gradient gives the direction of steepest decrease of the objective.

Q4. If f″(x) > 0 at a critical point, the point is:

Maximum Minimum
Saddle Inflection
Click here to view the answer

Correct option: Minimum

Positive second derivative implies local convexity, hence a local minimum.

Once you understand how objective functions are optimized, you’re ready to see how these ideas come together in real Machine Learning algorithms that learn patterns from data.

4. Machine Learning

Machine Learning KEKEW
Source: Giphy

Machine Learning focuses on algorithms that learn patterns from data, covering supervised and unsupervised methods, model evaluation, and the trade-off between bias and variance.

Articles:

Video learning: To reinforce concepts like overfitting, regularization, and distance-based learning, use the following YouTube playlist: Machine Learning

Questions (click to expand)

Q1. Which algorithm is most sensitive to feature scaling?

Decision Tree K-Nearest Neighbors
Naive Bayes Random Forest
Click here to view the answer

Correct option: K-Nearest Neighbors

KNN uses distances, so changing feature scales changes the distances and neighbors.

Q2. Ridge regression primarily addresses:

Bias Multicollinearity
Underfitting Class imbalance
Click here to view the answer

Correct option: Multicollinearity

L2 regularization stabilizes coefficients when predictors are correlated.

Q3. PCA reduces dimensionality by:

Maximizing variance Minimizing variance
Maximizing error Random projection
Click here to view the answer

Correct option: Maximizing variance

Principal components capture directions of maximum variance in the data.

Q4. Bias-variance trade-off refers to:

Model speed vs accuracy Underfitting vs overfitting
Training vs testing data Linear vs non-linear models
Click here to view the answer

Correct option: Underfitting vs overfitting

Higher model complexity tends to reduce bias but increase variance.

Having seen how models are trained and evaluated, the final step is understanding how Artificial Intelligence systems reason, search, and make decisions under uncertainty.

5. Artificial Intelligence

This is 100% AI
Source: Giphy

Artificial Intelligence deals with decision-making and reasoning, including search, logic, and probabilistic inference, enabling systems to act intelligently under uncertainty.

Articles:

Video learning: For visual walkthroughs of search algorithms, game-playing strategies, and inference methods, use the following YouTube playlist: Artificial Intelligence

Questions (click to expand)

Q1. BFS is preferred over DFS when:

Memory is limited Shortest path is required
Graph is deep Cycles exist
Click here to view the answer

Correct option: Shortest path is required

BFS guarantees the shortest path in unweighted graphs.

Q2. Minimax algorithm is used in:

Supervised learning Adversarial search
Clustering Reinforcement learning only
Click here to view the answer

Correct option: Adversarial search

Minimax models optimal play in two-player zero-sum games.

Q3. Conditional independence is crucial for:

Naive Bayes k-Means
PCA Linear Regression
Click here to view the answer

Correct option: Naive Bayes

Naive Bayes assumes features are conditionally independent given the class.

Q4. Variable elimination is an example of:

Approximate inference Exact inference
Sampling Heuristic search
Click here to view the answer

Correct option: Exact inference

Variable elimination computes exact marginals in probabilistic graphical models.

More help

To tell whether you are prepared on the subject, the questions would serve as a litmus test. If you struggled to get through the questions, then more learning is required. Here are all the YouTube playlists subject wise:

  1. Probability and Statistics
  2. Linear Algebra
  3. Calculus and Optimization
  4. Machine Learning
  5. Artificial Intelligence

If this learning material is too much for you, then you might consider short form content covering Artificial Intelligence and Data Science. 

If you were unable to find the resources helpful, then checkout the GitHub repository on GATE DA. Curated by aspirants who had cracked the exam, the repo is a treasure trove of content for data science and artificial intelligence.

GATE DA GitHub Repository
Source: GitHub

With the resources and the questions out of the way, the only thing left is for you to decide how you’re gonna approach the learning. 

I specialize in reviewing and refining AI-driven research, technical documentation, and content related to emerging AI technologies. My experience spans AI model training, data analysis, and information retrieval, allowing me to craft content that is both technically accurate and accessible.

Login to continue reading and enjoy expert-curated content.

Responses From Readers

Clear