datasets
🤗 The largest hub of ready-to-use datasets for ML models with fast, easy-to-use and efficient data manipulation tools
Top Related Projects
TFDS is a collection of datasets ready to use with TensorFlow, Jax, ...
A PyTorch repo for data loading and utilities to be shared by the PyTorch domain libraries.
Flexible and powerful data analysis / manipulation library for Python, providing labeled data structures similar to R data.frame objects, statistical functions, and much more
Parallel computing with task scheduling
scikit-learn: machine learning in Python
Quick Overview
The huggingface/datasets repository is a library that provides easy access to a wide range of datasets for machine learning tasks. It offers a simple and unified API to download, preprocess, and use datasets across various domains, including natural language processing, computer vision, and audio processing.
Pros
- Extensive collection of datasets from diverse domains
- Easy-to-use API for loading and preprocessing data
- Seamless integration with popular machine learning libraries like PyTorch and TensorFlow
- Supports streaming and memory-efficient data loading for large datasets
Cons
- Some datasets may have licensing restrictions or require additional authentication
- Limited support for custom dataset formats or structures
- Occasional issues with dataset versioning and reproducibility
- Dependency on external sources for dataset availability
Code Examples
Loading a dataset:
from datasets import load_dataset
dataset = load_dataset("glue", "mrpc")
print(dataset["train"][0])
Preprocessing a dataset:
from datasets import load_dataset
from transformers import AutoTokenizer
dataset = load_dataset("imdb")
tokenizer = AutoTokenizer.from_pretrained("bert-base-uncased")
def tokenize_function(examples):
return tokenizer(examples["text"], padding="max_length", truncation=True)
tokenized_dataset = dataset.map(tokenize_function, batched=True)
Streaming a large dataset:
from datasets import load_dataset
dataset = load_dataset("c4", "en", streaming=True)
for example in dataset["train"].take(5):
print(example["text"][:100])
Getting Started
To get started with the huggingface/datasets library, follow these steps:
- Install the library:
pip install datasets
- Load a dataset:
from datasets import load_dataset
# Load a specific dataset
dataset = load_dataset("glue", "mrpc")
# Access the data
print(dataset["train"][0])
# Get dataset info
print(dataset)
- Explore and preprocess the data:
# Get column names
print(dataset["train"].column_names)
# Apply a preprocessing function
def uppercase_text(example):
return {"text": example["text"].upper()}
processed_dataset = dataset["train"].map(uppercase_text)
print(processed_dataset[0])
Competitor Comparisons
TFDS is a collection of datasets ready to use with TensorFlow, Jax, ...
Pros of datasets (TensorFlow)
- Tightly integrated with TensorFlow ecosystem
- Optimized for TensorFlow-based workflows
- Includes TensorFlow-specific features like tf.data API compatibility
Cons of datasets (TensorFlow)
- Limited to TensorFlow-centric applications
- Smaller collection of datasets compared to Hugging Face
- Less flexibility for use with other ML frameworks
Code Comparison
datasets (Hugging Face):
from datasets import load_dataset
dataset = load_dataset("glue", "mrpc")
train_dataset = dataset["train"]
datasets (TensorFlow):
import tensorflow_datasets as tfds
dataset = tfds.load("glue/mrpc", split="train")
dataset = dataset.shuffle(1000).batch(32)
Key Differences
- Hugging Face datasets offers a more framework-agnostic approach
- TensorFlow datasets is more focused on TensorFlow-specific optimizations
- Hugging Face provides a larger variety of datasets across different domains
- TensorFlow datasets integrates seamlessly with TensorFlow's data pipeline
Both libraries aim to simplify dataset loading and preprocessing for machine learning tasks, but they cater to different use cases and preferences within the ML community.
A PyTorch repo for data loading and utilities to be shared by the PyTorch domain libraries.
Pros of data
- Tighter integration with PyTorch ecosystem
- More flexible data loading and processing pipelines
- Better support for custom datasets and data formats
Cons of data
- Smaller community and fewer pre-built datasets
- Less standardized dataset interfaces
- Steeper learning curve for beginners
Code Comparison
datasets:
from datasets import load_dataset
dataset = load_dataset("mnist")
train_data = dataset["train"]
data:
from torchvision.datasets import MNIST
from torch.utils.data import DataLoader
dataset = MNIST(root="data", train=True, download=True)
train_loader = DataLoader(dataset, batch_size=32, shuffle=True)
Summary
datasets offers a vast collection of ready-to-use datasets and a simpler interface, making it ideal for quick experimentation and research. data provides more flexibility and customization options, better suited for complex data processing pipelines and integration with PyTorch models. datasets is more beginner-friendly, while data offers greater control for advanced users. The choice between the two depends on the specific project requirements and the user's familiarity with the PyTorch ecosystem.
Flexible and powerful data analysis / manipulation library for Python, providing labeled data structures similar to R data.frame objects, statistical functions, and much more
Pros of pandas
- Mature and widely-used library with extensive documentation and community support
- Powerful data manipulation and analysis capabilities for structured data
- Seamless integration with other scientific Python libraries (NumPy, Matplotlib, etc.)
Cons of pandas
- Limited support for large-scale datasets and distributed computing
- Not specifically designed for machine learning tasks or handling diverse data types
- Steeper learning curve for beginners compared to more specialized libraries
Code Comparison
pandas:
import pandas as pd
df = pd.read_csv('data.csv')
filtered_df = df[df['column'] > 5]
result = filtered_df.groupby('category').mean()
datasets:
from datasets import load_dataset
dataset = load_dataset('csv', data_files='data.csv')
filtered_dataset = dataset.filter(lambda x: x['column'] > 5)
result = filtered_dataset.map(lambda x: {'mean': sum(x['values']) / len(x['values'])})
The pandas code demonstrates its strength in data manipulation and analysis, while the datasets code showcases its simplicity in loading and processing datasets for machine learning tasks. datasets is more focused on handling diverse data types and preparing data for ML models, whereas pandas excels in general-purpose data analysis and manipulation.
Parallel computing with task scheduling
Pros of Dask
- Designed for parallel and distributed computing, allowing for processing of larger-than-memory datasets
- Integrates well with existing Python ecosystems like NumPy and Pandas
- Offers flexible scheduling options for various computing environments
Cons of Dask
- Steeper learning curve compared to Datasets, especially for complex distributed computing tasks
- Less focus on machine learning and NLP-specific datasets
- May require more setup and configuration for distributed environments
Code Comparison
Datasets:
from datasets import load_dataset
dataset = load_dataset("mnist")
print(dataset["train"][0])
Dask:
import dask.dataframe as dd
df = dd.read_csv("large_dataset.csv")
result = df.groupby("column").mean().compute()
Summary
Datasets is tailored for machine learning and NLP tasks, offering easy access to pre-processed datasets. Dask, on the other hand, excels in parallel and distributed computing for large-scale data processing across various domains. While Datasets provides a simpler interface for working with ML datasets, Dask offers more flexibility and scalability for general data processing tasks.
scikit-learn: machine learning in Python
Pros of scikit-learn
- Comprehensive suite of machine learning algorithms and tools
- Well-established, mature library with extensive documentation
- Seamless integration with other scientific Python libraries (NumPy, SciPy, Pandas)
Cons of scikit-learn
- Limited support for deep learning and neural networks
- Not optimized for large-scale distributed computing
- Primarily focused on tabular data, less suitable for unstructured data like text or images
Code Comparison
scikit-learn:
from sklearn.model_selection import train_test_split
from sklearn.ensemble import RandomForestClassifier
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2)
clf = RandomForestClassifier()
clf.fit(X_train, y_train)
datasets:
from datasets import load_dataset
dataset = load_dataset("glue", "mrpc")
train_data = dataset["train"]
test_data = dataset["test"]
The scikit-learn example demonstrates data splitting and model training, while the datasets example shows loading a pre-defined dataset. scikit-learn focuses on machine learning algorithms, whereas datasets emphasizes data loading and preprocessing for various AI tasks, particularly in natural language processing.
Convert designs to code with AI
Introducing Visual Copilot: A new AI model to turn Figma designs to high quality code using your components.
Try Visual CopilotREADME
ð¤ Datasets is a lightweight library providing two main features:
- one-line dataloaders for many public datasets: one-liners to download and pre-process any of the major public datasets (image datasets, audio datasets, text datasets in 467 languages and dialects, etc.) provided on the HuggingFace Datasets Hub. With a simple command like
squad_dataset = load_dataset("squad")
, get any of these datasets ready to use in a dataloader for training/evaluating a ML model (Numpy/Pandas/PyTorch/TensorFlow/JAX), - efficient data pre-processing: simple, fast and reproducible data pre-processing for the public datasets as well as your own local datasets in CSV, JSON, text, PNG, JPEG, WAV, MP3, Parquet, etc. With simple commands like
processed_dataset = dataset.map(process_example)
, efficiently prepare the dataset for inspection and ML model evaluation and training.
ð Documentation ð Find a dataset in the Hub ð Share a dataset on the Hub
ð¤ Datasets is designed to let the community easily add and share new datasets.
ð¤ Datasets has many additional interesting features:
- Thrive on large datasets: ð¤ Datasets naturally frees the user from RAM memory limitation, all datasets are memory-mapped using an efficient zero-serialization cost backend (Apache Arrow).
- Smart caching: never wait for your data to process several times.
- Lightweight and fast with a transparent and pythonic API (multi-processing/caching/memory-mapping).
- Built-in interoperability with NumPy, pandas, PyTorch, TensorFlow 2 and JAX.
- Native support for audio and image data.
- Enable streaming mode to save disk space and start iterating over the dataset immediately.
ð¤ Datasets originated from a fork of the awesome TensorFlow Datasets and the HuggingFace team want to deeply thank the TensorFlow Datasets team for building this amazing library. More details on the differences between ð¤ Datasets and tfds
can be found in the section Main differences between ð¤ Datasets and tfds
.
Installation
With pip
ð¤ Datasets can be installed from PyPi and has to be installed in a virtual environment (venv or conda for instance)
pip install datasets
With conda
ð¤ Datasets can be installed using conda as follows:
conda install -c huggingface -c conda-forge datasets
Follow the installation pages of TensorFlow and PyTorch to see how to install them with conda.
For more details on installation, check the installation page in the documentation: https://huggingface.co/docs/datasets/installation
Installation to use with PyTorch/TensorFlow/pandas
If you plan to use ð¤ Datasets with PyTorch (1.0+), TensorFlow (2.2+) or pandas, you should also install PyTorch, TensorFlow or pandas.
For more details on using the library with NumPy, pandas, PyTorch or TensorFlow, check the quick start page in the documentation: https://huggingface.co/docs/datasets/quickstart
Usage
ð¤ Datasets is made to be very simple to use - the API is centered around a single function, datasets.load_dataset(dataset_name, **kwargs)
, that instantiates a dataset.
This library can be used for text/image/audio/etc. datasets. Here is an example to load a text dataset:
Here is a quick example:
from datasets import load_dataset
# Print all the available datasets
from huggingface_hub import list_datasets
print([dataset.id for dataset in list_datasets()])
# Load a dataset and print the first example in the training set
squad_dataset = load_dataset('squad')
print(squad_dataset['train'][0])
# Process the dataset - add a column with the length of the context texts
dataset_with_length = squad_dataset.map(lambda x: {"length": len(x["context"])})
# Process the dataset - tokenize the context texts (using a tokenizer from the ð¤ Transformers library)
from transformers import AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained('bert-base-cased')
tokenized_dataset = squad_dataset.map(lambda x: tokenizer(x['context']), batched=True)
If your dataset is bigger than your disk or if you don't want to wait to download the data, you can use streaming:
# If you want to use the dataset immediately and efficiently stream the data as you iterate over the dataset
image_dataset = load_dataset('cifar100', streaming=True)
for example in image_dataset["train"]:
break
For more details on using the library, check the quick start page in the documentation: https://huggingface.co/docs/datasets/quickstart and the specific pages on:
- Loading a dataset: https://huggingface.co/docs/datasets/loading
- What's in a Dataset: https://huggingface.co/docs/datasets/access
- Processing data with ð¤ Datasets: https://huggingface.co/docs/datasets/process
- Processing audio data: https://huggingface.co/docs/datasets/audio_process
- Processing image data: https://huggingface.co/docs/datasets/image_process
- Processing text data: https://huggingface.co/docs/datasets/nlp_process
- Streaming a dataset: https://huggingface.co/docs/datasets/stream
- Writing your own dataset loading script: https://huggingface.co/docs/datasets/dataset_script
- etc.
Add a new dataset to the Hub
We have a very detailed step-by-step guide to add a new dataset to the datasets already provided on the HuggingFace Datasets Hub.
You can find:
- how to upload a dataset to the Hub using your web browser or Python and also
- how to upload it using Git.
Main differences between ð¤ Datasets and tfds
If you are familiar with the great TensorFlow Datasets, here are the main differences between ð¤ Datasets and tfds
:
- the scripts in ð¤ Datasets are not provided within the library but are queried, downloaded/cached and dynamically loaded upon request
- the backend serialization of ð¤ Datasets is based on Apache Arrow instead of TF Records and leverage python dataclasses for info and features with some diverging features (we mostly don't do encoding and store the raw data as much as possible in the backend serialization cache).
- the user-facing dataset object of ð¤ Datasets is not a
tf.data.Dataset
but a built-in framework-agnostic dataset class with methods inspired by what we like intf.data
(like amap()
method). It basically wraps a memory-mapped Arrow table cache.
Disclaimers
ð¤ Datasets may run Python code defined by the dataset authors to parse certain data formats or structures. For security reasons, we ask users to:
- check the dataset scripts they're going to run beforehand and
- pin the
revision
of the repositories they use.
If you're a dataset owner and wish to update any part of it (description, citation, license, etc.), or do not want your dataset to be included in the Hugging Face Hub, please get in touch by opening a discussion or a pull request in the Community tab of the dataset page. Thanks for your contribution to the ML community!
BibTeX
If you want to cite our ð¤ Datasets library, you can use our paper:
@inproceedings{lhoest-etal-2021-datasets,
title = "Datasets: A Community Library for Natural Language Processing",
author = "Lhoest, Quentin and
Villanova del Moral, Albert and
Jernite, Yacine and
Thakur, Abhishek and
von Platen, Patrick and
Patil, Suraj and
Chaumond, Julien and
Drame, Mariama and
Plu, Julien and
Tunstall, Lewis and
Davison, Joe and
{\v{S}}a{\v{s}}ko, Mario and
Chhablani, Gunjan and
Malik, Bhavitvya and
Brandeis, Simon and
Le Scao, Teven and
Sanh, Victor and
Xu, Canwen and
Patry, Nicolas and
McMillan-Major, Angelina and
Schmid, Philipp and
Gugger, Sylvain and
Delangue, Cl{\'e}ment and
Matussi{\`e}re, Th{\'e}o and
Debut, Lysandre and
Bekman, Stas and
Cistac, Pierric and
Goehringer, Thibault and
Mustar, Victor and
Lagunas, Fran{\c{c}}ois and
Rush, Alexander and
Wolf, Thomas",
booktitle = "Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing: System Demonstrations",
month = nov,
year = "2021",
address = "Online and Punta Cana, Dominican Republic",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.emnlp-demo.21",
pages = "175--184",
abstract = "The scale, variety, and quantity of publicly-available NLP datasets has grown rapidly as researchers propose new tasks, larger models, and novel benchmarks. Datasets is a community library for contemporary NLP designed to support this ecosystem. Datasets aims to standardize end-user interfaces, versioning, and documentation, while providing a lightweight front-end that behaves similarly for small datasets as for internet-scale corpora. The design of the library incorporates a distributed, community-driven approach to adding datasets and documenting usage. After a year of development, the library now includes more than 650 unique datasets, has more than 250 contributors, and has helped support a variety of novel cross-dataset research projects and shared tasks. The library is available at https://github.com/huggingface/datasets.",
eprint={2109.02846},
archivePrefix={arXiv},
primaryClass={cs.CL},
}
If you need to cite a specific version of our ð¤ Datasets library for reproducibility, you can use the corresponding version Zenodo DOI from this list.
Top Related Projects
TFDS is a collection of datasets ready to use with TensorFlow, Jax, ...
A PyTorch repo for data loading and utilities to be shared by the PyTorch domain libraries.
Flexible and powerful data analysis / manipulation library for Python, providing labeled data structures similar to R data.frame objects, statistical functions, and much more
Parallel computing with task scheduling
scikit-learn: machine learning in Python
Convert designs to code with AI
Introducing Visual Copilot: A new AI model to turn Figma designs to high quality code using your components.
Try Visual Copilot