25 Open Datasets for Deep Learning Every Data Scientist Must Work With

Pranav Dar 31 May, 2024
11 min read

Introduction

Deep learning is a subset of machine learning based on neural networks with representation learning. The key to mastering this topic (or most fields in life) is practice. There are a variety of practice problems available in deep learning, ranging from image processing to speech recognition. But where can you get the sample datasets for these practice problems? In this article, we have listed a collection of openly available high-quality datasets for deep learning enthusiasts. We have also added a few practice problems towards the end of this article, for you to use these public datasets.

25 Open Datasets for Deep Learning Every Data Scientist Must Work With

The Need for Open-Source Datasets

Open source datasets are much needed for data science students, researchers, and working professionals to test out various artificial intelligence (AI) and machine learning (ML) algorithms. Problems such as time series forecasting, computer vision, regression, semantic analysis, data analysis, and more, require large datasets to work on.

Working on these datasets will make you a better data scientist and the amount of learning you will have will be invaluable in your career. This article also includes papers with state-of-the-art (SOTA) results for you to go through and improve your models.

A lot of research papers you see these days use proprietary datasets that are usually not released to the general public. This becomes a problem, if you want to learn and apply your newly acquired skills. If you have faced this problem, we have a solution for you. Here’s a list of openly available datasets for your perusal.

Open Source Datasets for Deep Learning

How to Use These Datasets?

First things first – these datasets are huge in size! So make sure you have a fast internet connection with no / very high limit on the amount of data you can download.

There are numerous ways how you can use these datasets. You can use them to apply various deep learning techniques. You can use them to hone your skills, understand how to identify and structure each problem, think of unique use cases, and publish your findings for everyone to see!

In this article, we have included 25 versatile datasets you can use for deep learning problems. The datasets are divided into three categories – Image Processing, Natural Language Processing, and Audio/Speech Processing.

Let’s dive into it!

Image Datasets for Deep Learning

1. MNIST

MNIST | 25 Open Datasets for Deep Learning

MNIST is one of the most popular deep learning datasets out there. It’s a dataset of handwritten digits and contains a training set of 60,000 examples and a test set of 10,000 examples. It’s a good database for trying learning techniques and deep recognition patterns on real-world data while spending minimum time and effort in data preprocessing.

Size: ~50 MB

Number of Records: 70,000 images in 10 classes

SOTADynamic Routing Between Capsules

2. MS-COCO

MS-COCO | 25 Open Datasets for Deep Learning

COCO is a large-scale and rich for object detection, segmentation and captioning dataset. It has several features:

  • Object segmentation
  • Recognition in context
  • Superpixel stuff segmentation
  • 330K images (>200K labeled)
  • 1.5 million object instances
  • 80 object categories
  • 91 stuff categories
  • 5 captions per image
  • 250,000 people with keypoints

Size: ~25 GB (Compressed)

Number of Records: 330K images, 80 object categories, 5 captions per image, 250,000 people with key points

SOTAMask R-CNN

Bored with Datasets? Solve real life project on Deep Learning 

3. ImageNet

ImageNet | 25 Open Datasets for Deep Learning

ImageNet is a dataset of images that are organized according to the WordNet hierarchy. WordNet contains approximately 100,000 phrases and ImageNet has provided around 1000 images on average to illustrate each phrase.

Size: ~150GB

Number of Records: Total number of images: ~1,500,000; each with multiple bounding boxes and respective class labels

SOTAAggregated Residual Transformations for Deep Neural Networks

4. Open Images Dataset

Open Images Dataset

Open Images is a dataset of almost 9 million URLs for images. These images have been annotated with image-level labels bounding boxes spanning thousands of classes. The dataset contains a training set of 9,011,219 images, a validation set of 41,260 images, and a test set of 125,436 images.

Size: 500 GB (Compressed)

Number of Records: 9,011,219 images with more than 5k labels

SOTA: Resnet 101 image classification model (trained on V2 data): Model checkpoint, Checkpoint readme, Inference code.

5. VisualQA

VisualQA | VQA

VQA is a dataset containing open-ended questions about images. These questions require an understanding of vision and language. Some of the interesting features of this dataset are:

  • 265,016 images (COCO and abstract scenes)
  • At least 3 questions (5.4 questions on average) per image
  • 10 ground truth answers per question
  • 3 plausible (but likely incorrect) answers per question
  • Automatic evaluation metric

Size: 25 GB (Compressed)

Number of Records: 265,016 images, at least 3 questions per image, 10 ground truth answers per question

SOTATips and Tricks for Visual Question Answering: Learnings from the 2017 Challenge

6. The Street View House Numbers (SVHN)

Street View House Numbers (SVHN)

This is a real-world image dataset for developing object detection algorithms. This requires minimum data preprocessing. It is similar to the MNIST dataset mentioned in this list but has more labeled data (over 600,000 labeled images). The data in the SVHN dataset has been collected from house numbers viewed in Google Street View.

Size: 2.5 GB

Number of Records: 6,30,420 images in 10 classes

SOTADistributional Smoothing With Virtual Adversarial Training

7. CIFAR-10

CIFAR-10

This dataset is another one for image classification. It consists of 60,000 images of 10 classes (each class is represented as a row in the above image). In total, there are 50,000 training images and 10,000 test images. The CIFAR-10 dataset is divided into 6 parts – 5 training batches and 1 test batch. Each batch has 10,000 images.

Size: 170 MB

Number of Records: 60,000 images in 10 classes

SOTAShakeDrop regularization

8. Fashion-MNIST

Fashion-MNIST | 25 Open Datasets for Deep Learning

Fashion-MNIST consists of 60,000 training images and 10,000 test images. It is an MNIST-like fashion product database. The developers believe MNIST has been overused so they created this as a direct replacement for that dataset. Each image is in greyscale and associated with a label from 10 classes.

Size: 30 MB

Number of Records: 70,000 images in 10 classes

SOTARandom Erasing Data Augmentation

Natural Language Processing (NLP) Datasets

9. IMDB Reviews

This is a dream dataset for movie lovers. It is meant for binary sentiment classification and has far more data than any previous datasets in this field. Apart from the training and test review examples, there is further unlabeled data for use as well. Raw text and preprocessed bag of word formats have also been included.

Size: 80 MB

Number of Records: 25,000 highly polar movie reviews for training, and 25,000 for testing

SOTALearning Structured Text Representations

10. Twenty Newsgroups

This dataset, as the name suggests, contains information about newsgroups. To curate this dataset, 1000 Usenet articles were taken from 20 different newsgroups. The articles have typical features like subject lines, signatures, and quotes.

Size: 20 MB

Number of Records: 20,000 messages taken from 20 newsgroups

SOTAVery Deep Convolutional Networks for Text Classification

11. Sentiment140

Sentiment140 is a dataset that can be used for sentiment analysis. A popular dataset, it is perfect to start off your NLP journey. Emotions have already been removed from the data. The final dataset has the below 6 features:

  • polarity of the tweet
  • id of the tweet
  • date of the tweet
  • the query
  • username of the tweeter
  • text of the tweet

Size: 80 MB (Compressed)

Number of Records: 1,60,000 tweets

SOTAAssessing State-of-the-Art Sentiment Models on State-of-the-Art Sentiment Datasets

12. WordNet

As mentioned in the ImageNet dataset above, WordNet is a large database of English synsets. Synsets are groups of synonyms that each describe a different concept. WordNet’s structure makes it a very useful tool for NLP.

Size: 10 MB

Number of Records: 117,000 synsets is linked to other synsets by means of a small number of “conceptual relations.

SOTA: Wordnets: State of the Art and Perspectives

13. Yelp Reviews

This is an open dataset released by Yelp for learning purposes. It consists of millions of user reviews, businesses attributes, and over 200,000 pictures from multiple metropolitan areas. This is a very commonly used dataset for NLP challenges globally.

Size: 2.66 GB JSON, 2.9 GB SQL and 7.5 GB Photos (all compressed)

Number of Records: 5,200,000 reviews, 174,000 business attributes, 200,000 pictures and 11 metropolitan areas

SOTAAttentive Convolution

14. The Wikipedia Corpus

This dataset is a collection of all the text on Wikipedia. It contains almost 1.9 billion words from more than 4 million articles. What makes this a powerful NLP dataset is that you search by word, phrase or part of a paragraph itself.

Size: 20 MB

Number of Records: 4,400,000 articles containing 1.9 billion words

SOTABreaking The Softmax Bottelneck: A High-Rank RNN language Model

15. The Blog Authorship Corpus

This dataset consists of blog posts collected from thousands of bloggers and has been gathered from blogger.com. Each blog is provided as a separate file. Each blog contains a minimum of 200 occurrences of commonly used English words.

Size: 300 MB

Number of Records: 681,288 posts with over 140 million words

SOTACharacter-level and Multi-channel Convolutional Neural Networks for Large-scale Authorship Attribution

16. Machine Translation of Various Languages

This dataset consists of training data for four European languages. The task here is to improve the current translation methods. You can participate in any of the following language pairs:

  • English-Chinese and Chinese-English
  • English-Czech and Czech-English
  • English-Estonian and Estonian-English
  • English-Finnish and Finnish-English
  • English-German and German-English
  • English-Kazakh and Kazakh-English
  • English-Russian and Russian-English
  • English-Turkish and Turkish-English

Size: ~15 GB

Number of Records: ~30,000,000 sentences and their translations

SOTAAttention Is All You Need

Engage with real-life projects on Natural Language Processing here.

Audio/Speech Datasets

17. Free Spoken Digit Dataset

Another entry in this list inspired by the MNIST dataset! This one was created to solve the task of identifying spoken digits in audio samples. It’s an open dataset so the hope is that it will keep growing as people keep contributing more samples. Currently, it contains the following characteristics:

  • 3 speakers
  • 1,500 recordings (50 of each digit per speaker)
  • English pronunciations

Size: 10 MB

Number of Records: 1,500 audio samples

SOTARaw Waveform-based Audio Classification Using Sample-level CNN Architectures

18. Free Music Archive (FMA)

FMA is a dataset for music analysis. The dataset consists of full-length and HQ audio, pre-computed features, and track and user-level metadata. It is an open dataset created to evaluate several tasks in MIR. Below is the list of CSV files the dataset has along with what they include:

  • tracks.csv: per track metadata such as ID, title, artist, genres, tags, and play counts, for all 106,574 tracks.
  • genres.csv: all 163 genre IDs with their name and parent (used to infer the genre hierarchy and top-level genres).
  • features.csv: common features extracted with librosa.
  • echonest.csv: audio features provided by Echonest (now Spotify) for a subset of 13,129 tracks.

Size: ~1000 GB

Number of Records: ~100,000 tracks

SOTALearning to Recognize Musical Genre from Audio

19. Ballroom

This dataset contains ballroom dancing audio files. A few characteristic excerpts of many dance styles are provided in real audio format. Below are a few characteristics of the dataset:

  • Total number of instances: 698
  • Duration: ~30 s
  • Total duration: ~20940 s

Size: 14GB (Compressed)

Number of Records: ~700 audio samples

SOTAA Multi-Model Approach To Beat Tracking Considering Heterogeneous Music Styles

20. Million Song Dataset

Million Song Dataset

The Million Song Dataset is a freely available collection of audio features and metadata for a million contemporary popular music tracks. Its purposes are:

  • To encourage research on algorithms that scale to commercial sizes
  • To provide a reference dataset for evaluating research
  • As a shortcut alternative to creating a large dataset with APIs (e.g. The Echo Nest’s)
  • To help new researchers get started in the MIR field

The core of the dataset is the feature analysis and metadata for one million songs. The dataset does not include any audio, only the derived features. The sample audio can be fetched from services like 7digital, using code provided by Columbia University.

Size: 280 GB

Number of Records: PS – its a million songs!

SOTA: Preliminary Study on a Recommender System for the Million Songs Dataset Challenge

21. LibriSpeech

This dataset is a large-scale corpus of around 1000 hours of English speech. The data has been sourced from audiobooks from the LibriVox project. It has been segmented and aligned properly. If you’re looking for a starting point, check out already prepared Acoustic models that are trained on this data set at kaldi-asr.org and language models, suitable for evaluation, at http://www.openslr.org/11/.

Size: ~60 GB

Number of Records: 1000 hours of speech

SOTALetter-Based Speech Recognition with Gated ConvNets

22. VoxCeleb

VoxCeleb is a large-scale speaker identification dataset. It contains around 100,000 utterances by 1,251 celebrities, extracted from YouTube videos. The data is mostly gender balanced (males comprise of 55%). The celebrities span a diverse range of accents, professions and age. There is no overlap between the development and test sets. It’s an intriguing use case for isolating and identifying which superstar the voice belongs to.

Size: 150 MB

Number of Records: 100,000 utterances by 1,251 celebrities

SOTA: VoxCeleb: a large-scale speaker identification dataset

Analytics Vidhya Practice Problems

For your practice, we also provide real-life problems and datasets to get your hands dirty. In this section, we’ve listed down the deep learning practice problems on our DataHack platform.

23. X Sentiment Analysis

Hate Speech in the form of racism and sexism has become a nuisance on X (formerly, Twitter) and it is important to segregate these sorts of tweets from the rest. In this practice problem, we provide Twitter data that has both normal and hate tweets. Your task as a data scientist is to identify the tweets that are hate tweets and those that are not.

Size: 3 MB

Number of Records: 31,962 tweets

24. Age Detection of Indian Actors

This is a fascinating challenge for any deep learning enthusiast. The dataset contains thousands of images of Indian actors and your task is to identify their age. All the images are manually selected and cropped from the video frames resulting in a high degree of variability interms of scale, pose, expression, illumination, age, resolution, occlusion, and makeup.

Size: 48 MB (Compressed)

Number of Records: 19,906 images in the training set and 6636 in the test set

SOTA: Hands on with Deep Learning – Solution for Age Detection Practice Problem

25. Urban Sound Classification

This dataset consists of more than 8000 sound excerpts of urban sounds from 10 classes. This practice problem is meant to introduce you to audio processing in the usual classification scenario.

Size: Training set – 3 GB (Compressed), Test set – 2 GB (Compressed)

Number of Records: 8732 labeled sound excerpts (<=4s) of urban sounds from 10 classes

Conclusion

Mastering deep learning requires practice, and having access to the right datasets can make a huge difference in your learning journey. With the rise of open-source models, we now have access to a number of training datasets. However, these new datasets may be specific to each of those models, letting us test, experiment, and build on them.

Each dataset comes with specific characteristics and benchmarks that can help you test and improve your models. Whether you’re a student, researcher, or professional, these resources offer valuable opportunities to apply and enhance your skills in real-world scenarios.

You can use these public datasets to apply various deep learning algorithms and improve your skillset. Do try out the practice problems listed in this article, and let us know in the comments if the datasets were helpful in your attempts.

Frequently Asked Questions

Q1. What are datasets in deep learning?

A. Datasets are collections of data that are used to train, validate, and test models. In deep learning, these datasets are essential for developing and evaluating algorithms. Deep learning datasets can contain data in various forms, such as images, text, audio, and video.

Q2. Where can I get ML datasets?

A. This article is a comprehensive resource for open-source datasets. You can find more open datasets for machine learning on Kaggle, GitHub, UCI Machine Learning Repository, Amazon’s Registry of Open Data, and Google’s Datasets Search Engine.

Q3. What is a good dataset size for deep learning?

A. The ideal dataset size for deep learning depends on the complexity of the task and the model architecture being used. Generally, larger datasets tend to yield better performance. A good rule of thumb is to aim for thousands to millions of data points for effective training. However, it’s also important to balance the dataset size with computational resources and model capacity to prevent overfitting.

Pranav Dar 31 May, 2024

Senior Editor at Analytics Vidhya. Data visualization practitioner who loves reading and delving deeper into the data science and machine learning arts. Always looking for new ways to improve processes using ML and AI.

Frequently Asked Questions

Lorem ipsum dolor sit amet, consectetur adipiscing elit,

Responses From Readers

Clear

Sujatha Sivaraman
Sujatha Sivaraman 29 Mar, 2018

Thanks Pranav. This is indeed going to be useful to strengthen your practice

Sunny Toms
Sunny Toms 29 Mar, 2018

Good work Pranav. This is good info for deep learning self learners. Thank you.

Fredrik
Fredrik 29 Mar, 2018

Very good, thanks!

Mithlesh Patel
Mithlesh Patel 29 Mar, 2018

Thankyou Pranav, very useful for self learners.

Raghu S
Raghu S 29 Mar, 2018

Sir, where can I get medical image dataset. Please provide necessary information.

N
N 29 Mar, 2018

Great job Pranav, very usefuil

Manish
Manish 29 Mar, 2018

Great job Pranav. Keep it.

Swapna
Swapna 29 Mar, 2018

Thanks Pranav. Good info for self learners

Mathias Müller
Mathias Müller 30 Mar, 2018

A machine translation researcher here, Regarding the machine translation data set you present: You clearly have not researched this well. You link randomly to the WMT machine translation shared task of 2011, although there are 7 newer editions, including 2018 (http://statmt.org/wmt18/index.html). Also, the "Attention is All you Need" paper from Google is _not_ state of the art on the 2011 data you are describing, the paper only has results from the 2014 shared task), and only from English to German, and English to French. By the way, the data sets are of course not called "Machine Translation of European Languages". WMT data is comprised of several individual data sets, such as Europarl, JRC Aquis, News Commentary or OpenSubtitles. Please do your homework.

Peter
Peter 31 Mar, 2018

Thank You!

ahmed awaad
ahmed awaad 01 Apr, 2018

we need dataset about link quality

David
David 02 Apr, 2018

We maintain this huge list of color-names: https://github.com/meodai/color-names/ Could also be useful to categorise or name colors

V V CHAKRADHAR
V V CHAKRADHAR 02 Apr, 2018

Hey Pranav , I am working on Resume Ranking . Can you help me with a good link !!

Darryl Ricketts
Darryl Ricketts 04 Apr, 2018

The Wiki Corpus you link to is pay; $245 for the complete dump.

Morten Bjoernsvik
Morten Bjoernsvik 07 Apr, 2018

I prefer rest apis so I do not have to download: https://frost.met.no/ - 15K met stations around the globe the oldest data I've found is from 1937 https://data.cityofnewyork.us/browse - I wich my city had the same more than 1000 datasets updated daily.

Prashnna
Prashnna 18 Apr, 2018

Those of whom are interested in the Devnagari dataset for character recognition can utilize this (https://github.com/Prasanna1991/DHCD_Dataset) resource. It's bigger than MNIST. P.S. Devnagari script is famous in Indian Subcontinent. (Nepal, India, Bangladesh, etc.)

Akhila
Akhila 02 May, 2018

Hi pranav, i am a research student. i need to write the state of art about Automatic image orientation Detection using Convolutional Neural Networks, so suggest me some list of recently updated datasets. it would be a great help for me.

Sachin
Sachin 06 May, 2018

Thanks for the useful information. Can you please provide the dataset for link prediction.

Vishaal
Vishaal 08 May, 2018

Thanks, this is a great resource!

Leo
Leo 28 Jul, 2018

Bookmarked. Thanks a lot! One note: Wikipedia corpus is definitely bigger than 30MB :)

Aleksandra
Aleksandra 27 Sep, 2018

Thanks for good sources of data. I am looking for a set with a coloured full body shot of a person. Could you suggest any?

Kavita Lalwani
Kavita Lalwani 14 Nov, 2018

Thanks for the great resource, lots of info here. Would love to connect. Do check out our training platform at Experfy — www.experfy.com/training We are a Harvard-incubated company and the biggest data science/AI/machine learning marketplace for enterprises. Hope to connect. Cheers, Kavita Lalwani

Kunal jain
Kunal jain 22 Nov, 2018

Hi, thank you for the given information....

Patricia
Patricia 27 Jun, 2022

This is a great blog post - So clear and easy to follow. All your hard work is so much appreciated.