Top Related Projects
TensorFlow documentation
Keras documentation, hosted live at keras.io
12 weeks, 26 lessons, 52 quizzes, classic Machine Learning for all
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
scikit-learn: machine learning in Python
The fastai deep learning library
Quick Overview
PyTorch Tutorials is an official repository containing a comprehensive set of tutorials and examples for PyTorch, a popular open-source machine learning library. These tutorials cover a wide range of topics, from basic tensor operations to advanced deep learning techniques, providing both beginners and experienced practitioners with valuable resources to learn and master PyTorch.
Pros
- Extensive coverage of PyTorch features and applications
- Well-structured and regularly updated content
- Official documentation, ensuring accuracy and reliability
- Includes both text explanations and interactive Jupyter notebooks
Cons
- Some tutorials may be overwhelming for absolute beginners
- Occasional inconsistencies in coding style across different tutorials
- Limited coverage of certain specialized topics or cutting-edge techniques
Code Examples
- Basic tensor operations:
import torch
# Create a tensor
x = torch.tensor([1, 2, 3])
# Perform operations
y = x + 2
z = torch.matmul(x, y)
print(y)
print(z)
- Simple neural network:
import torch
import torch.nn as nn
class SimpleNet(nn.Module):
def __init__(self):
super(SimpleNet, self).__init__()
self.fc1 = nn.Linear(10, 5)
self.fc2 = nn.Linear(5, 2)
def forward(self, x):
x = torch.relu(self.fc1(x))
x = self.fc2(x)
return x
model = SimpleNet()
print(model)
- Training loop example:
import torch.optim as optim
criterion = nn.CrossEntropyLoss()
optimizer = optim.SGD(model.parameters(), lr=0.01)
for epoch in range(10):
for inputs, labels in dataloader:
optimizer.zero_grad()
outputs = model(inputs)
loss = criterion(outputs, labels)
loss.backward()
optimizer.step()
Getting Started
To get started with PyTorch Tutorials:
-
Clone the repository:
git clone https://github.com/pytorch/tutorials.git
-
Install dependencies:
pip install -r requirements.txt
-
Navigate to a specific tutorial directory and run the Jupyter notebook:
jupyter notebook
-
Alternatively, you can view the tutorials online at the official PyTorch website: https://pytorch.org/tutorials/
Competitor Comparisons
TensorFlow documentation
Pros of docs
- More comprehensive documentation covering a wider range of TensorFlow features
- Better organization with clear categorization of topics
- Includes API reference alongside tutorials and guides
Cons of docs
- Less focus on practical, hands-on examples compared to tutorials
- May be overwhelming for beginners due to the sheer volume of information
- Slower to update with the latest features and changes
Code Comparison
tutorials:
import torch
x = torch.rand(5, 3)
print(x)
docs:
import tensorflow as tf
x = tf.random.uniform((5, 3))
print(x)
Both repositories provide code examples, but tutorials tends to offer more complete, runnable examples, while docs often provides shorter snippets to illustrate specific concepts or API usage.
Summary
While docs offers a more comprehensive and well-organized resource for TensorFlow, tutorials provides a more hands-on, practical approach to learning PyTorch. The choice between the two depends on the user's preferred learning style and the specific framework they're working with.
Keras documentation, hosted live at keras.io
Pros of keras-io
- More comprehensive documentation with detailed guides and examples
- Better organized structure with clear categorization of topics
- Includes interactive Colab notebooks for hands-on learning
Cons of keras-io
- Less frequent updates compared to pytorch/tutorials
- Focuses primarily on Keras, limiting exposure to other deep learning frameworks
- May not cover some advanced topics found in pytorch/tutorials
Code Comparison
Keras-io example:
model = keras.Sequential([
keras.layers.Dense(64, activation='relu'),
keras.layers.Dense(10, activation='softmax')
])
model.compile(optimizer='adam', loss='categorical_crossentropy')
PyTorch Tutorials example:
class Net(nn.Module):
def __init__(self):
super(Net, self).__init__()
self.fc1 = nn.Linear(784, 64)
self.fc2 = nn.Linear(64, 10)
def forward(self, x):
x = F.relu(self.fc1(x))
return F.softmax(self.fc2(x), dim=1)
Both repositories offer valuable resources for learning their respective frameworks. Keras-io provides a more structured and comprehensive learning experience, while PyTorch Tutorials offers a wider range of advanced topics and more frequent updates. The code examples demonstrate the different approaches to model creation in Keras and PyTorch.
12 weeks, 26 lessons, 52 quizzes, classic Machine Learning for all
Pros of ML-For-Beginners
- Comprehensive curriculum covering various ML topics
- Beginner-friendly with clear explanations and hands-on projects
- Language-agnostic approach, not tied to a specific framework
Cons of ML-For-Beginners
- Less focus on deep learning compared to PyTorch Tutorials
- May not cover advanced topics in as much depth
- Fewer code examples specific to production-ready implementations
Code Comparison
ML-For-Beginners (using scikit-learn):
from sklearn.model_selection import train_test_split
from sklearn.linear_model import LogisticRegression
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2)
model = LogisticRegression()
model.fit(X_train, y_train)
PyTorch Tutorials:
import torch
import torch.nn as nn
model = nn.Linear(input_size, output_size)
criterion = nn.MSELoss()
optimizer = torch.optim.SGD(model.parameters(), lr=0.01)
ML-For-Beginners provides a broader introduction to machine learning concepts using various tools, while PyTorch Tutorials focuses specifically on deep learning with PyTorch. The former is more suitable for beginners looking to understand ML fundamentals, while the latter is better for those wanting to dive deep into PyTorch and neural networks.
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Pros of transformers
- Extensive library of pre-trained models for various NLP tasks
- High-level APIs for easy model fine-tuning and deployment
- Active community and frequent updates
Cons of transformers
- Steeper learning curve for beginners
- More focused on NLP tasks, less general-purpose than tutorials
Code Comparison
transformers:
from transformers import BertTokenizer, BertForSequenceClassification
tokenizer = BertTokenizer.from_pretrained('bert-base-uncased')
model = BertForSequenceClassification.from_pretrained('bert-base-uncased')
inputs = tokenizer("Hello, my dog is cute", return_tensors="pt")
outputs = model(**inputs)
tutorials:
import torch
import torch.nn as nn
class SimpleModel(nn.Module):
def __init__(self):
super(SimpleModel, self).__init__()
self.linear = nn.Linear(10, 1)
def forward(self, x):
return self.linear(x)
The transformers example showcases the ease of using pre-trained models, while the tutorials example demonstrates basic PyTorch model creation. transformers provides a higher-level abstraction, whereas tutorials offers more flexibility for custom architectures.
scikit-learn: machine learning in Python
Pros of scikit-learn
- Comprehensive library for traditional machine learning algorithms
- Easier to use for beginners and non-deep learning tasks
- Extensive documentation and community support
Cons of scikit-learn
- Limited support for deep learning and neural networks
- Less flexibility for custom model architectures
- Not optimized for GPU acceleration
Code Comparison
scikit-learn:
from sklearn.ensemble import RandomForestClassifier
from sklearn.datasets import make_classification
X, y = make_classification(n_samples=1000, n_features=4)
clf = RandomForestClassifier()
clf.fit(X, y)
PyTorch:
import torch
import torch.nn as nn
class SimpleNet(nn.Module):
def __init__(self):
super(SimpleNet, self).__init__()
self.fc = nn.Linear(4, 2)
def forward(self, x):
return self.fc(x)
model = SimpleNet()
The scikit-learn example shows a simple random forest classifier, while the PyTorch example demonstrates a basic neural network structure. scikit-learn's code is more concise for traditional ML tasks, whereas PyTorch offers more flexibility for deep learning architectures.
The fastai deep learning library
Pros of fastai
- Provides a higher-level API, making it easier for beginners to get started with deep learning
- Includes built-in best practices and optimized defaults for common tasks
- Offers a comprehensive library of pre-built models and techniques
Cons of fastai
- Less flexibility for customizing low-level components compared to PyTorch
- Steeper learning curve for understanding fastai-specific concepts and abstractions
- Smaller community and ecosystem compared to PyTorch
Code Comparison
fastai:
from fastai.vision.all import *
path = untar_data(URLs.PETS)
dls = ImageDataLoaders.from_folder(path, valid_pct=0.2, seed=42)
learn = cnn_learner(dls, resnet34, metrics=error_rate)
learn.fine_tune(5)
PyTorch tutorials:
import torch
import torchvision
model = torchvision.models.resnet34(pretrained=True)
criterion = nn.CrossEntropyLoss()
optimizer = torch.optim.SGD(model.parameters(), lr=0.001, momentum=0.9)
for epoch in range(5):
# Training loop implementation
The fastai code is more concise and abstracts away many details, while the PyTorch tutorials code provides more explicit control over the training process.
Convert designs to code with AI
Introducing Visual Copilot: A new AI model to turn Figma designs to high quality code using your components.
Try Visual CopilotREADME
PyTorch Tutorials
All the tutorials are now presented as sphinx style documentation at:
https://pytorch.org/tutorials
Asking a question
If you have a question about a tutorial, post in https://dev-discuss.pytorch.org/ rather than creating an issue in this repo. Your question will be answered much faster on the dev-discuss forum.
Submitting an issue
You can submit the following types of issues:
- Feature request - request a new tutorial to be added. Please explain why this tutorial is needed and how it demonstrates PyTorch value.
- Bug report - report a failure or outdated information in an existing tutorial. When submitting a bug report, please run:
python3 -m torch.utils.collect_env
to get information about your environment and add the output to the bug report.
Contributing
We use sphinx-gallery's notebook styled examples to create the tutorials. Syntax is very simple. In essence, you write a slightly well formatted Python file and it shows up as an HTML page. In addition, a Jupyter notebook is autogenerated and available to run in Google Colab.
Here is how you can create a new tutorial (for a detailed description, see CONTRIBUTING.md):
NOTE: Before submitting a new tutorial, read PyTorch Tutorial Submission Policy.
- Create a Python file. If you want it executed while inserted into documentation, save the file with the suffix
tutorial
so that the file name isyour_tutorial.py
. - Put it in one of the
beginner_source
,intermediate_source
,advanced_source
directory based on the level of difficulty. If it is a recipe, add it torecipes_source
. For tutorials demonstrating unstable prototype features, add to theprototype_source
. - For Tutorials (except if it is a prototype feature), include it in the
toctree
directive and create acustomcarditem
in index.rst. - For Tutorials (except if it is a prototype feature), create a thumbnail in the index.rst file using a command like
.. customcarditem:: beginner/your_tutorial.html
. For Recipes, create a thumbnail in the recipes_index.rst
If you are starting off with a Jupyter notebook, you can use this script to convert the notebook to Python file. After conversion and addition to the project, please make sure that section headings and other things are in logical order.
Building locally
The tutorial build is very large and requires a GPU. If your machine does not have a GPU device, you can preview your HTML build without actually downloading the data and running the tutorial code:
- Install required dependencies by running:
pip install -r requirements.txt
.
Typically, you would run either in
conda
orvirtualenv
. If you want to usevirtualenv
, in the root of the repo, run:virtualenv venv
, thensource venv/bin/activate
.
- If you have a GPU-powered laptop, you can build using
make docs
. This will download the data, execute the tutorials and build the documentation todocs/
directory. This might take about 60-120 min for systems with GPUs. If you do not have a GPU installed on your system, then see next step. - You can skip the computationally intensive graph generation by running
make html-noplot
to build basic html documentation to_build/html
. This way, you can quickly preview your tutorial.
Building a single tutorial
You can build a single tutorial by using the GALLERY_PATTERN
environment variable. For example to run only neural_style_transfer_tutorial.py
, run:
GALLERY_PATTERN="neural_style_transfer_tutorial.py" make html
or
GALLERY_PATTERN="neural_style_transfer_tutorial.py" sphinx-build . _build
The GALLERY_PATTERN
variable respects regular expressions.
About contributing to PyTorch Documentation and Tutorials
- You can find information about contributing to PyTorch documentation in the PyTorch Repo README.md file.
- Additional information can be found in PyTorch CONTRIBUTING.md.
License
PyTorch Tutorials is BSD licensed, as found in the LICENSE file.
Top Related Projects
TensorFlow documentation
Keras documentation, hosted live at keras.io
12 weeks, 26 lessons, 52 quizzes, classic Machine Learning for all
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
scikit-learn: machine learning in Python
The fastai deep learning library
Convert designs to code with AI
Introducing Visual Copilot: A new AI model to turn Figma designs to high quality code using your components.
Try Visual Copilot