Convert Figma logo to code with AI

huggingface logodatasets

🤗 The largest hub of ready-to-use datasets for ML models with fast, easy-to-use and efficient data manipulation tools

19,154
2,668
19,154
769

Top Related Projects

TFDS is a collection of datasets ready to use with TensorFlow, Jax, ...

1,115

A PyTorch repo for data loading and utilities to be shared by the PyTorch domain libraries.

43,524

Flexible and powerful data analysis / manipulation library for Python, providing labeled data structures similar to R data.frame objects, statistical functions, and much more

12,495

Parallel computing with task scheduling

scikit-learn: machine learning in Python

Quick Overview

The huggingface/datasets repository is a library that provides easy access to a wide range of datasets for machine learning tasks. It offers a simple and unified API to download, preprocess, and use datasets across various domains, including natural language processing, computer vision, and audio processing.

Pros

  • Extensive collection of datasets from diverse domains
  • Easy-to-use API for loading and preprocessing data
  • Seamless integration with popular machine learning libraries like PyTorch and TensorFlow
  • Supports streaming and memory-efficient data loading for large datasets

Cons

  • Some datasets may have licensing restrictions or require additional authentication
  • Limited support for custom dataset formats or structures
  • Occasional issues with dataset versioning and reproducibility
  • Dependency on external sources for dataset availability

Code Examples

Loading a dataset:

from datasets import load_dataset

dataset = load_dataset("glue", "mrpc")
print(dataset["train"][0])

Preprocessing a dataset:

from datasets import load_dataset
from transformers import AutoTokenizer

dataset = load_dataset("imdb")
tokenizer = AutoTokenizer.from_pretrained("bert-base-uncased")

def tokenize_function(examples):
    return tokenizer(examples["text"], padding="max_length", truncation=True)

tokenized_dataset = dataset.map(tokenize_function, batched=True)

Streaming a large dataset:

from datasets import load_dataset

dataset = load_dataset("c4", "en", streaming=True)
for example in dataset["train"].take(5):
    print(example["text"][:100])

Getting Started

To get started with the huggingface/datasets library, follow these steps:

  1. Install the library:
pip install datasets
  1. Load a dataset:
from datasets import load_dataset

# Load a specific dataset
dataset = load_dataset("glue", "mrpc")

# Access the data
print(dataset["train"][0])

# Get dataset info
print(dataset)
  1. Explore and preprocess the data:
# Get column names
print(dataset["train"].column_names)

# Apply a preprocessing function
def uppercase_text(example):
    return {"text": example["text"].upper()}

processed_dataset = dataset["train"].map(uppercase_text)
print(processed_dataset[0])

Competitor Comparisons

TFDS is a collection of datasets ready to use with TensorFlow, Jax, ...

Pros of datasets (TensorFlow)

  • Tightly integrated with TensorFlow ecosystem
  • Optimized for TensorFlow-based workflows
  • Includes TensorFlow-specific features like tf.data API compatibility

Cons of datasets (TensorFlow)

  • Limited to TensorFlow-centric applications
  • Smaller collection of datasets compared to Hugging Face
  • Less flexibility for use with other ML frameworks

Code Comparison

datasets (Hugging Face):

from datasets import load_dataset

dataset = load_dataset("glue", "mrpc")
train_dataset = dataset["train"]

datasets (TensorFlow):

import tensorflow_datasets as tfds

dataset = tfds.load("glue/mrpc", split="train")
dataset = dataset.shuffle(1000).batch(32)

Key Differences

  • Hugging Face datasets offers a more framework-agnostic approach
  • TensorFlow datasets is more focused on TensorFlow-specific optimizations
  • Hugging Face provides a larger variety of datasets across different domains
  • TensorFlow datasets integrates seamlessly with TensorFlow's data pipeline

Both libraries aim to simplify dataset loading and preprocessing for machine learning tasks, but they cater to different use cases and preferences within the ML community.

1,115

A PyTorch repo for data loading and utilities to be shared by the PyTorch domain libraries.

Pros of data

  • Tighter integration with PyTorch ecosystem
  • More flexible data loading and processing pipelines
  • Better support for custom datasets and data formats

Cons of data

  • Smaller community and fewer pre-built datasets
  • Less standardized dataset interfaces
  • Steeper learning curve for beginners

Code Comparison

datasets:

from datasets import load_dataset

dataset = load_dataset("mnist")
train_data = dataset["train"]

data:

from torchvision.datasets import MNIST
from torch.utils.data import DataLoader

dataset = MNIST(root="data", train=True, download=True)
train_loader = DataLoader(dataset, batch_size=32, shuffle=True)

Summary

datasets offers a vast collection of ready-to-use datasets and a simpler interface, making it ideal for quick experimentation and research. data provides more flexibility and customization options, better suited for complex data processing pipelines and integration with PyTorch models. datasets is more beginner-friendly, while data offers greater control for advanced users. The choice between the two depends on the specific project requirements and the user's familiarity with the PyTorch ecosystem.

43,524

Flexible and powerful data analysis / manipulation library for Python, providing labeled data structures similar to R data.frame objects, statistical functions, and much more

Pros of pandas

  • Mature and widely-used library with extensive documentation and community support
  • Powerful data manipulation and analysis capabilities for structured data
  • Seamless integration with other scientific Python libraries (NumPy, Matplotlib, etc.)

Cons of pandas

  • Limited support for large-scale datasets and distributed computing
  • Not specifically designed for machine learning tasks or handling diverse data types
  • Steeper learning curve for beginners compared to more specialized libraries

Code Comparison

pandas:

import pandas as pd

df = pd.read_csv('data.csv')
filtered_df = df[df['column'] > 5]
result = filtered_df.groupby('category').mean()

datasets:

from datasets import load_dataset

dataset = load_dataset('csv', data_files='data.csv')
filtered_dataset = dataset.filter(lambda x: x['column'] > 5)
result = filtered_dataset.map(lambda x: {'mean': sum(x['values']) / len(x['values'])})

The pandas code demonstrates its strength in data manipulation and analysis, while the datasets code showcases its simplicity in loading and processing datasets for machine learning tasks. datasets is more focused on handling diverse data types and preparing data for ML models, whereas pandas excels in general-purpose data analysis and manipulation.

12,495

Parallel computing with task scheduling

Pros of Dask

  • Designed for parallel and distributed computing, allowing for processing of larger-than-memory datasets
  • Integrates well with existing Python ecosystems like NumPy and Pandas
  • Offers flexible scheduling options for various computing environments

Cons of Dask

  • Steeper learning curve compared to Datasets, especially for complex distributed computing tasks
  • Less focus on machine learning and NLP-specific datasets
  • May require more setup and configuration for distributed environments

Code Comparison

Datasets:

from datasets import load_dataset

dataset = load_dataset("mnist")
print(dataset["train"][0])

Dask:

import dask.dataframe as dd

df = dd.read_csv("large_dataset.csv")
result = df.groupby("column").mean().compute()

Summary

Datasets is tailored for machine learning and NLP tasks, offering easy access to pre-processed datasets. Dask, on the other hand, excels in parallel and distributed computing for large-scale data processing across various domains. While Datasets provides a simpler interface for working with ML datasets, Dask offers more flexibility and scalability for general data processing tasks.

scikit-learn: machine learning in Python

Pros of scikit-learn

  • Comprehensive suite of machine learning algorithms and tools
  • Well-established, mature library with extensive documentation
  • Seamless integration with other scientific Python libraries (NumPy, SciPy, Pandas)

Cons of scikit-learn

  • Limited support for deep learning and neural networks
  • Not optimized for large-scale distributed computing
  • Primarily focused on tabular data, less suitable for unstructured data like text or images

Code Comparison

scikit-learn:

from sklearn.model_selection import train_test_split
from sklearn.ensemble import RandomForestClassifier
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2)
clf = RandomForestClassifier()
clf.fit(X_train, y_train)

datasets:

from datasets import load_dataset
dataset = load_dataset("glue", "mrpc")
train_data = dataset["train"]
test_data = dataset["test"]

The scikit-learn example demonstrates data splitting and model training, while the datasets example shows loading a pre-defined dataset. scikit-learn focuses on machine learning algorithms, whereas datasets emphasizes data loading and preprocessing for various AI tasks, particularly in natural language processing.

Convert Figma logo designs to code with AI

Visual Copilot

Introducing Visual Copilot: A new AI model to turn Figma designs to high quality code using your components.

Try Visual Copilot

README

Hugging Face Datasets Library

Build GitHub Documentation GitHub release Number of datasets Contributor Covenant DOI

🤗 Datasets is a lightweight library providing two main features:

  • one-line dataloaders for many public datasets: one-liners to download and pre-process any of the number of datasets major public datasets (image datasets, audio datasets, text datasets in 467 languages and dialects, etc.) provided on the HuggingFace Datasets Hub. With a simple command like squad_dataset = load_dataset("squad"), get any of these datasets ready to use in a dataloader for training/evaluating a ML model (Numpy/Pandas/PyTorch/TensorFlow/JAX),
  • efficient data pre-processing: simple, fast and reproducible data pre-processing for the public datasets as well as your own local datasets in CSV, JSON, text, PNG, JPEG, WAV, MP3, Parquet, etc. With simple commands like processed_dataset = dataset.map(process_example), efficiently prepare the dataset for inspection and ML model evaluation and training.

🎓 Documentation 🔎 Find a dataset in the Hub 🌟 Share a dataset on the Hub

🤗 Datasets is designed to let the community easily add and share new datasets.

🤗 Datasets has many additional interesting features:

  • Thrive on large datasets: 🤗 Datasets naturally frees the user from RAM memory limitation, all datasets are memory-mapped using an efficient zero-serialization cost backend (Apache Arrow).
  • Smart caching: never wait for your data to process several times.
  • Lightweight and fast with a transparent and pythonic API (multi-processing/caching/memory-mapping).
  • Built-in interoperability with NumPy, pandas, PyTorch, TensorFlow 2 and JAX.
  • Native support for audio and image data.
  • Enable streaming mode to save disk space and start iterating over the dataset immediately.

🤗 Datasets originated from a fork of the awesome TensorFlow Datasets and the HuggingFace team want to deeply thank the TensorFlow Datasets team for building this amazing library. More details on the differences between 🤗 Datasets and tfds can be found in the section Main differences between 🤗 Datasets and tfds.

Installation

With pip

🤗 Datasets can be installed from PyPi and has to be installed in a virtual environment (venv or conda for instance)

pip install datasets

With conda

🤗 Datasets can be installed using conda as follows:

conda install -c huggingface -c conda-forge datasets

Follow the installation pages of TensorFlow and PyTorch to see how to install them with conda.

For more details on installation, check the installation page in the documentation: https://huggingface.co/docs/datasets/installation

Installation to use with PyTorch/TensorFlow/pandas

If you plan to use 🤗 Datasets with PyTorch (1.0+), TensorFlow (2.2+) or pandas, you should also install PyTorch, TensorFlow or pandas.

For more details on using the library with NumPy, pandas, PyTorch or TensorFlow, check the quick start page in the documentation: https://huggingface.co/docs/datasets/quickstart

Usage

🤗 Datasets is made to be very simple to use - the API is centered around a single function, datasets.load_dataset(dataset_name, **kwargs), that instantiates a dataset.

This library can be used for text/image/audio/etc. datasets. Here is an example to load a text dataset:

Here is a quick example:

from datasets import load_dataset

# Print all the available datasets
from huggingface_hub import list_datasets
print([dataset.id for dataset in list_datasets()])

# Load a dataset and print the first example in the training set
squad_dataset = load_dataset('squad')
print(squad_dataset['train'][0])

# Process the dataset - add a column with the length of the context texts
dataset_with_length = squad_dataset.map(lambda x: {"length": len(x["context"])})

# Process the dataset - tokenize the context texts (using a tokenizer from the 🤗 Transformers library)
from transformers import AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained('bert-base-cased')

tokenized_dataset = squad_dataset.map(lambda x: tokenizer(x['context']), batched=True)

If your dataset is bigger than your disk or if you don't want to wait to download the data, you can use streaming:

# If you want to use the dataset immediately and efficiently stream the data as you iterate over the dataset
image_dataset = load_dataset('cifar100', streaming=True)
for example in image_dataset["train"]:
    break

For more details on using the library, check the quick start page in the documentation: https://huggingface.co/docs/datasets/quickstart and the specific pages on:

Add a new dataset to the Hub

We have a very detailed step-by-step guide to add a new dataset to the number of datasets datasets already provided on the HuggingFace Datasets Hub.

You can find:

Main differences between 🤗 Datasets and tfds

If you are familiar with the great TensorFlow Datasets, here are the main differences between 🤗 Datasets and tfds:

  • the scripts in 🤗 Datasets are not provided within the library but are queried, downloaded/cached and dynamically loaded upon request
  • the backend serialization of 🤗 Datasets is based on Apache Arrow instead of TF Records and leverage python dataclasses for info and features with some diverging features (we mostly don't do encoding and store the raw data as much as possible in the backend serialization cache).
  • the user-facing dataset object of 🤗 Datasets is not a tf.data.Dataset but a built-in framework-agnostic dataset class with methods inspired by what we like in tf.data (like a map() method). It basically wraps a memory-mapped Arrow table cache.

Disclaimers

🤗 Datasets may run Python code defined by the dataset authors to parse certain data formats or structures. For security reasons, we ask users to:

  • check the dataset scripts they're going to run beforehand and
  • pin the revision of the repositories they use.

If you're a dataset owner and wish to update any part of it (description, citation, license, etc.), or do not want your dataset to be included in the Hugging Face Hub, please get in touch by opening a discussion or a pull request in the Community tab of the dataset page. Thanks for your contribution to the ML community!

BibTeX

If you want to cite our 🤗 Datasets library, you can use our paper:

@inproceedings{lhoest-etal-2021-datasets,
    title = "Datasets: A Community Library for Natural Language Processing",
    author = "Lhoest, Quentin  and
      Villanova del Moral, Albert  and
      Jernite, Yacine  and
      Thakur, Abhishek  and
      von Platen, Patrick  and
      Patil, Suraj  and
      Chaumond, Julien  and
      Drame, Mariama  and
      Plu, Julien  and
      Tunstall, Lewis  and
      Davison, Joe  and
      {\v{S}}a{\v{s}}ko, Mario  and
      Chhablani, Gunjan  and
      Malik, Bhavitvya  and
      Brandeis, Simon  and
      Le Scao, Teven  and
      Sanh, Victor  and
      Xu, Canwen  and
      Patry, Nicolas  and
      McMillan-Major, Angelina  and
      Schmid, Philipp  and
      Gugger, Sylvain  and
      Delangue, Cl{\'e}ment  and
      Matussi{\`e}re, Th{\'e}o  and
      Debut, Lysandre  and
      Bekman, Stas  and
      Cistac, Pierric  and
      Goehringer, Thibault  and
      Mustar, Victor  and
      Lagunas, Fran{\c{c}}ois  and
      Rush, Alexander  and
      Wolf, Thomas",
    booktitle = "Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing: System Demonstrations",
    month = nov,
    year = "2021",
    address = "Online and Punta Cana, Dominican Republic",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2021.emnlp-demo.21",
    pages = "175--184",
    abstract = "The scale, variety, and quantity of publicly-available NLP datasets has grown rapidly as researchers propose new tasks, larger models, and novel benchmarks. Datasets is a community library for contemporary NLP designed to support this ecosystem. Datasets aims to standardize end-user interfaces, versioning, and documentation, while providing a lightweight front-end that behaves similarly for small datasets as for internet-scale corpora. The design of the library incorporates a distributed, community-driven approach to adding datasets and documenting usage. After a year of development, the library now includes more than 650 unique datasets, has more than 250 contributors, and has helped support a variety of novel cross-dataset research projects and shared tasks. The library is available at https://github.com/huggingface/datasets.",
    eprint={2109.02846},
    archivePrefix={arXiv},
    primaryClass={cs.CL},
}

If you need to cite a specific version of our 🤗 Datasets library for reproducibility, you can use the corresponding version Zenodo DOI from this list.