Convert Figma logo to code with AI

RedditSota logostate-of-the-art-result-for-machine-learning-problems

This repository provides state of the art (SoTA) results for all machine learning problems. We do our best to keep this repository up to date. If you do find a problem's SoTA result is out of date or missing, please raise this as an issue or submit Google form (with this information: research paper name, dataset, metric, source code and year). We will fix it immediately.

8,948
1,314
8,948
14

Top Related Projects

30,928

A library for efficient similarity search and clustering of dense vectors.

🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

77,006

Models and examples built with TensorFlow

82,049

Tensors and Dynamic neural networks in Python with strong GPU acceleration

34,658

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

37,810

TensorFlow code and pre-trained models for BERT

Quick Overview

The RedditSota/state-of-the-art-result-for-machine-learning-problems repository is a curated collection of state-of-the-art (SOTA) results for various machine learning tasks. It serves as a comprehensive reference for researchers and practitioners, providing links to papers, code implementations, and datasets for different problem domains in machine learning.

Pros

  • Comprehensive coverage of various machine learning tasks and domains
  • Regularly updated with the latest SOTA results
  • Includes links to papers, code implementations, and datasets
  • Serves as a valuable resource for researchers and practitioners

Cons

  • Relies on community contributions, which may lead to inconsistent updates
  • Some links may become outdated over time
  • May not cover all niche or emerging areas of machine learning
  • Lacks detailed explanations or comparisons of different SOTA approaches

Code Examples

This repository is not a code library but rather a collection of links and information. Therefore, there are no code examples to provide.

Getting Started

As this is not a code library, there are no specific getting started instructions. However, users can navigate the repository by browsing the different sections organized by machine learning tasks and domains. To contribute or suggest updates, users can follow the repository's contribution guidelines and submit pull requests with new SOTA results or corrections to existing information.

Competitor Comparisons

30,928

A library for efficient similarity search and clustering of dense vectors.

Pros of faiss

  • Highly optimized C++ library for efficient similarity search and clustering of dense vectors
  • Supports GPU acceleration for faster processing of large-scale datasets
  • Provides a Python interface for easy integration with machine learning workflows

Cons of faiss

  • Focused solely on similarity search and clustering, not a comprehensive ML resource
  • Requires more technical expertise to implement and use effectively
  • Limited to dense vector representations, may not be suitable for all ML problems

Code comparison

faiss:

import faiss
import numpy as np

d = 64  # dimension
nb = 100000  # database size
nq = 10000  # nb of queries
xb = np.random.random((nb, d)).astype('float32')
xq = np.random.random((nq, d)).astype('float32')

index = faiss.IndexFlatL2(d)
index.add(xb)
D, I = index.search(xq, k=4)

state-of-the-art-result-for-machine-learning-problems:

# Image Classification

## CIFAR-10

* [96.53%] [Wide Residual Networks](http://arxiv.org/abs/1605.07146) by Sergey Zagoruyko, Nikos Komodakis

## CIFAR-100

* [82.95%] [Wide Residual Networks](http://arxiv.org/abs/1605.07146) by Sergey Zagoruyko, Nikos Komodakis

Summary

faiss is a specialized library for similarity search and clustering, while state-of-the-art-result-for-machine-learning-problems is a curated list of top-performing ML models across various tasks. faiss offers high-performance implementations but requires more technical expertise, whereas state-of-the-art-result-for-machine-learning-problems provides a comprehensive overview of ML advancements but lacks implementation details.

🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

Pros of transformers

  • Provides a comprehensive library of pre-trained models and tools for natural language processing tasks
  • Offers easy-to-use APIs for fine-tuning and deploying models
  • Actively maintained with frequent updates and community support

Cons of transformers

  • Focuses primarily on transformer-based models, limiting coverage of other machine learning approaches
  • Requires more computational resources due to the complexity of transformer models
  • May have a steeper learning curve for beginners compared to a curated list of results

Code comparison

transformers:

from transformers import AutoModelForSequenceClassification, AutoTokenizer

model = AutoModelForSequenceClassification.from_pretrained("bert-base-uncased")
tokenizer = AutoTokenizer.from_pretrained("bert-base-uncased")

state-of-the-art-result-for-machine-learning-problems:

## Image Classification

* [ImageNet](http://www.image-net.org/)
  * [ConvNeXt: A ConvNet for the 2020s](https://arxiv.org/abs/2201.03545) (2022)

The transformers repository provides actual code implementations, while state-of-the-art-result-for-machine-learning-problems offers a curated list of research papers and their results for various machine learning tasks.

77,006

Models and examples built with TensorFlow

Pros of models

  • Comprehensive collection of official TensorFlow implementations
  • Well-maintained with regular updates and contributions from Google researchers
  • Includes pre-trained models and detailed documentation for each implementation

Cons of models

  • Focused solely on TensorFlow, limiting its scope for other frameworks
  • May not always include the absolute latest state-of-the-art results
  • Requires more setup and understanding of TensorFlow to use effectively

Code Comparison

models:

import tensorflow as tf
from official.vision.image_classification import resnet_model
model = resnet_model.resnet50(num_classes=1000)

state-of-the-art-result-for-machine-learning-problems:

## Image Classification

* [ImageNet Classification](https://github.com/RedditSota/state-of-the-art-result-for-machine-learning-problems#imagenet-classification)
* [CIFAR-10 Classification](https://github.com/RedditSota/state-of-the-art-result-for-machine-learning-problems#cifar-10-classification)

Summary

models provides official TensorFlow implementations with detailed documentation and pre-trained models, making it ideal for TensorFlow users. state-of-the-art-result-for-machine-learning-problems offers a broader overview of state-of-the-art results across various frameworks and problems, serving as a comprehensive reference for researchers and practitioners. The choice between the two depends on whether you need specific TensorFlow implementations or a general overview of current best practices in machine learning.

82,049

Tensors and Dynamic neural networks in Python with strong GPU acceleration

Pros of pytorch

  • Comprehensive deep learning framework with extensive functionality
  • Large, active community providing support and contributions
  • Seamless integration with CUDA for GPU acceleration

Cons of pytorch

  • Steeper learning curve for beginners
  • Larger codebase and installation size
  • May be overkill for simple machine learning tasks

Code comparison

pytorch:

import torch

x = torch.tensor([1, 2, 3])
y = torch.tensor([4, 5, 6])
z = torch.matmul(x, y)

state-of-the-art-result-for-machine-learning-problems:

# Image Classification

## CIFAR-10

* [96.53%] [Wide Residual Networks](http://arxiv.org/abs/1605.07146) (2016)
* [96.43%] [Fractional Max-Pooling](http://arxiv.org/abs/1412.6071) (2015)

The state-of-the-art-result-for-machine-learning-problems repository is a curated list of top-performing models for various machine learning tasks, while pytorch is a full-fledged deep learning framework. The former provides a quick reference for researchers and practitioners to stay updated on the latest achievements, while the latter offers tools and libraries for implementing and training models. The code examples highlight this difference, with pytorch showing actual model implementation and state-of-the-art-result-for-machine-learning-problems presenting performance metrics in a markdown format.

34,658

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Pros of DeepSpeed

  • Focuses on optimizing and scaling deep learning training
  • Provides a comprehensive suite of optimization techniques
  • Actively maintained by Microsoft with frequent updates

Cons of DeepSpeed

  • More complex to set up and use compared to a curated list
  • Requires more technical knowledge to implement effectively
  • May not cover as wide a range of ML problems as a general SOTA list

Code Comparison

DeepSpeed:

import deepspeed
model_engine, optimizer, _, _ = deepspeed.initialize(
    args=args, model=model, model_parameters=params
)

state-of-the-art-result-for-machine-learning-problems:

# Image Classification on ImageNet

* [EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks](https://arxiv.org/abs/1905.11946) (2019)

Summary

DeepSpeed is a powerful tool for optimizing deep learning training, offering advanced techniques for scaling and efficiency. However, it requires more technical expertise to implement. The state-of-the-art-result-for-machine-learning-problems repository provides a curated list of SOTA results across various ML tasks, making it easier to reference current benchmarks but lacking the optimization capabilities of DeepSpeed.

37,810

TensorFlow code and pre-trained models for BERT

Pros of BERT

  • Focused on a specific, powerful NLP model with pre-trained weights
  • Includes detailed implementation and usage examples
  • Actively maintained by Google Research team

Cons of BERT

  • Limited to BERT model and its variants
  • Requires more computational resources to run and fine-tune
  • Steeper learning curve for beginners in NLP

Code Comparison

BERT example:

import tensorflow as tf
from bert import modeling

bert_config = modeling.BertConfig.from_json_file("bert_config.json")
model = modeling.BertModel(config=bert_config, is_training=True, input_ids=input_ids)

state-of-the-art-result-for-machine-learning-problems doesn't provide code examples, as it's primarily a curated list of SOTA results and papers.

Summary

BERT is a specialized repository focusing on a specific NLP model, offering implementation details and pre-trained weights. It's well-maintained but requires more resources and expertise to use effectively.

state-of-the-art-result-for-machine-learning-problems is a comprehensive list of SOTA results across various ML domains, providing a broader overview of the field but lacking specific implementations or code examples.

Choose BERT for deep dives into transformer-based NLP models, and state-of-the-art-result-for-machine-learning-problems for staying updated on the latest achievements across multiple ML areas.

Convert Figma logo designs to code with AI

Visual Copilot

Introducing Visual Copilot: A new AI model to turn Figma designs to high quality code using your components.

Try Visual Copilot

README

State-of-the-art result for all Machine Learning Problems

LAST UPDATE: 20th Februray 2019

NEWS: I am looking for a Collaborator esp who does research in NLP, Computer Vision and Reinforcement learning. If you are not a researcher, but you are willing, contact me. Email me: yxt.stoaml@gmail.com

This repository provides state-of-the-art (SoTA) results for all machine learning problems. We do our best to keep this repository up to date. If you do find a problem's SoTA result is out of date or missing, please raise this as an issue (with this information: research paper name, dataset, metric, source code and year). We will fix it immediately.

You can also submit this Google Form if you are new to Github.

This is an attempt to make one stop for all types of machine learning problems state of the art result. I can not do this alone. I need help from everyone. Please submit the Google form/raise an issue if you find SOTA result for a dataset. Please share this on Twitter, Facebook, and other social media.

This summary is categorized into:

Supervised Learning

NLP

1. Language Modelling

Research Paper Datasets Metric Source Code Year
Language Models are Unsupervised Multitask Learners
  • PTB
  • WikiText-2
  • Perplexity: 35.76
  • Perplexity: 18.34
Tensorflow 2019
BREAKING THE SOFTMAX BOTTLENECK: A HIGH-RANK RNN LANGUAGE MODEL
  • PTB
  • WikiText-2
  • Perplexity: 47.69
  • Perplexity: 40.68
Pytorch 2017
DYNAMIC EVALUATION OF NEURAL SEQUENCE MODELS
  • PTB
  • WikiText-2
  • Perplexity: 51.1
  • Perplexity: 44.3
Pytorch 2017
Averaged Stochastic Gradient Descent
with Weight Dropped LSTM or QRNN
  • PTB
  • WikiText-2
  • Perplexity: 52.8
  • Perplexity: 52.0
Pytorch 2017
FRATERNAL DROPOUT
  • PTB
  • WikiText-2
  • Perplexity: 56.8
  • Perplexity: 64.1
Pytorch 2017
Factorization tricks for LSTM networks One Billion Word Benchmark Perplexity: 23.36 Tensorflow 2017

2. Machine Translation

Research Paper Datasets Metric Source Code Year
Understanding Back-Translation at Scale
  • WMT 2014 English-to-French
  • WMT 2014 English-to-German
  • BLEU: 45.6
  • BLEU: 35.0
2018
WEIGHTED TRANSFORMER NETWORK FOR MACHINE TRANSLATION
  • WMT 2014 English-to-French
  • WMT 2014 English-to-German
  • BLEU: 41.4
  • BLEU: 28.9
2017
Attention Is All You Need
  • WMT 2014 English-to-French
  • WMT 2014 English-to-German
  • BLEU: 41.0
  • BLEU: 28.4
2017
NON-AUTOREGRESSIVE NEURAL MACHINE TRANSLATION
  • WMT16 Ro→En
  • BLEU: 31.44
2017
Improving Neural Machine Translation with Conditional Sequence Generative Adversarial Nets
  • NIST02
  • NIST03
  • NIST04
  • NIST05
  • 38.74
  • 36.01
  • 37.54
  • 33.76
  • 2017

    3. Text Classification

    Research Paper Datasets Metric Source Code Year
    Learning Structured Text Representations Yelp Accuracy: 68.6 2017
    Attentive Convolution Yelp Accuracy: 67.36 2017

    4. Natural Language Inference

    Leader board:

    Stanford Natural Language Inference (SNLI)

    MultiNLI

    Research Paper Datasets Metric Source Code Year
    NATURAL LANGUAGE INFERENCE OVER INTERACTION SPACE Stanford Natural Language Inference (SNLI) Accuracy: 88.9 Tensorflow 2017
    BERT-LARGE (ensemble) Multi-Genre Natural Language Inference (MNLI)
    • Matched accuracy: 86.7
    • Mismatched accuracy: 85.9
    2018

    5. Question Answering

    Leader Board

    SQuAD

    Research Paper Datasets Metric Source Code Year
    BERT-LARGE (ensemble) The Stanford Question Answering Dataset
    • Exact Match: 87.4
    • F1: 93.2
    2018

    6. Named entity recognition

    Research Paper Datasets Metric Source Code Year
    Named Entity Recognition in Twitter using Images and Text Ritter
    • F-measure: 0.59
    NOT FOUND 2017

    7. Abstractive Summarization

    Research PaperDatasetsMetricSource CodeYear
    Cutting-off redundant repeating generations
    for neural abstractive summarization
    • DUC-2004
    • Gigaword
    • DUC-2004
      • ROUGE-1: 32.28
      • ROUGE-2: 10.54
      • ROUGE-L: 27.80
    • Gigaword
      • ROUGE-1: 36.30
      • ROUGE-2: 17.31
      • ROUGE-L: 33.88
    NOT YET AVAILABLE2017
    Convolutional Sequence to Sequence
    • DUC-2004
    • Gigaword
    • DUC-2004
      • ROUGE-1: 33.44
      • ROUGE-2: 10.84
      • ROUGE-L: 26.90
    • Gigaword
      • ROUGE-1: 35.88
      • ROUGE-2: 27.48
      • ROUGE-L: 33.29
    PyTorch2017

    8. Dependency Parsing

    Research PaperDatasetsMetricSource CodeYear
    Globally Normalized Transition-Based Neural Networks
    • Final CoNLL ’09 dependency parsing
    • 94.08% UAS accurancy
    • 92.15% LAS accurancy
    • 2017

    Computer Vision

    1. Classification

               
    Research Paper Datasets Metric Source Code Year
    Dynamic Routing Between Capsules
    • MNIST
    • Test Error: 0.25±0.005
    2017
    High-Performance Neural Networks for Visual Object Classification
    • NORB
    • Test Error: 2.53 ± 0.40
    2011
    Giant AmoebaNet with GPipe
    • CIFAR-10
    • CIFAR-100
    • ImageNet-1k
    • ...
    • Test Error: 1.0%
    • Test Error: 8.7%
    • Top-1 Error 15.7
    • ...
    2018
    ShakeDrop regularization
    • CIFAR-10
    • CIFAR-100
    • Test Error: 2.31%
    • Test Error: 12.19%
    2017
    Aggregated Residual Transformations for Deep Neural Networks
    • CIFAR-10
    • Test Error: 3.58%
    2017
    Random Erasing Data Augmentation
    • CIFAR-10
    • CIFAR-100
    • Fashion-MNIST
    • Test Error: 3.08%
    • Test Error: 17.73%
    • Test Error: 3.65%
    Pytorch 2017
    EraseReLU: A Simple Way to Ease the Training of Deep Convolution Neural Networks
    • CIFAR-10
    • CIFAR-100
    • Test Error: 3.56%
    • Test Error: 16.53%
    Pytorch 2017
    Dynamic Routing Between Capsules
    • MultiMNIST
    • Test Error: 5%
    2017
    Learning Transferable Architectures for Scalable Image Recognition
    • ImageNet-1k
    • Top-1 Error:17.3
    2017
    Squeeze-and-Excitation Networks
    • ImageNet-1k
    • Top-1 Error: 18.68
    2017
    Aggregated Residual Transformations for Deep Neural Networks
    • ImageNet-1k
    • Top-1 Error: 20.4%
    2016

    2. Instance Segmentation

    Research Paper Datasets Metric Source Code Year
    Mask R-CNN
    • COCO
    • Average Precision: 37.1%
    2017

    3. Visual Question Answering

    Research Paper Datasets Metric Source Code Year
    Tips and Tricks for Visual Question Answering: Learnings from the 2017 Challenge
    • VQA
    • Overall score: 69
    2017

    4. Person Re-identification

         
    Research Paper Datasets Metric Source Code Year
    Random Erasing Data Augmentation
    • Rank-1: 89.13 mAP: 83.93
    • Rank-1: 84.02 mAP: 78.28
    • labeled (Rank-1: 63.93 mAP: 65.05) detected (Rank-1: 64.43 mAP: 64.75)
    Pytorch 2017

    Speech

    Speech SOTA

    1. ASR

    Research Paper Datasets Metric Source Code Year
    The Microsoft 2017 Conversational Speech Recognition System
    • Switchboard Hub5'00
    • WER: 5.1
    2017
    The CAPIO 2017 Conversational Speech Recognition System
    • Switchboard Hub5'00
    • WER: 5.0
    2017

    Semi-supervised Learning

    Computer Vision

         
    Research Paper Datasets Metric Source Code Year
    DISTRIBUTIONAL SMOOTHINGWITH VIRTUAL ADVERSARIAL TRAINING
    • SVHN
    • NORB
    • Test error: 24.63
    • Test error: 9.88
    Theano 2016
    Virtual Adversarial Training: a Regularization Method for Supervised and Semi-supervised Learning
    • MNIST
    • Test error: 1.27
    2017
    Few Shot Object Detection
    • VOC2007
    • VOC2012
    • mAP : 41.7
    • mAP : 35.4
    2017
    Unlabeled Samples Generated by GAN Improve the Person Re-identification Baseline in vitro
    • Rank-1: 83.97 mAP: 66.07
    • Rank-1: 84.6 mAP: 87.4
    • Rank-1: 67.68 mAP: 47.13
    •          
    • Test Accuracy: 84.4
    Matconvnet 2017

    Unsupervised Learning

    Computer Vision

    1. Generative Model
    Research Paper Datasets Metric Source Code Year
    PROGRESSIVE GROWING OF GANS FOR IMPROVED QUALITY, STABILITY, AND VARIATION Unsupervised CIFAR 10 Inception score: 8.80 Theano 2017

    NLP

    Machine Translation

    Research Paper Datasets Metric Source Code Year
    UNSUPERVISED MACHINE TRANSLATION USING MONOLINGUAL CORPORA ONLY
    • Multi30k-Task1(en-fr fr-en de-en en-de)
    • BLEU:(32.76 32.07 26.26 22.74)
    2017
    Unsupervised Neural Machine Translation with Weight Sharing
    • WMT14(en-fr fr-en)
    • WMT16 (de-en en-de)
    • BLEU:(16.97 15.58)
    • BLEU:(14.62 10.86)
    2018

    Transfer Learning

    Research Paper Datasets Metric Source Code Year
    One Model To Learn Them All
    • WMT EN → DE
    • WMT EN → FR (BLEU)
    • ImageNet (top-5 accuracy)
    • BLEU: 21.2
    • BLEU:30.5
    • 86%
    2017

    Reinforcement Learning

    Research Paper Datasets Metric Source Code Year
    Mastering the game of Go without human knowledge the game of Go ElO Rating: 5185 2017

    Email: yxt.stoaml@gmail.com