horovod
Distributed training framework for TensorFlow, Keras, PyTorch, and Apache MXNet.
Top Related Projects
An Open Source Machine Learning Framework for Everyone
Tensors and Dynamic neural networks in Python with strong GPU acceleration
Scalable, Portable and Distributed Gradient Boosting (GBDT, GBRT or GBM) Library, for Python, R, Java, Scala, C++ and more. Runs on single machine, Hadoop, Spark, Dask, Flink and DataFlow
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Open source platform for the machine learning lifecycle
Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
Quick Overview
Horovod is a distributed deep learning training framework for TensorFlow, Keras, PyTorch, and Apache MXNet. It was developed by Uber and is designed to make distributed deep learning fast and easy to use. Horovod extends single-GPU training to multiple GPUs across multiple nodes, enabling efficient scaling of deep learning models.
Pros
- Easy integration with popular deep learning frameworks
- Efficient distributed training with minimal code changes
- Supports both CPU and GPU training
- Includes advanced features like gradient compression and mixed-precision training
Cons
- Requires additional setup and configuration for distributed environments
- May have a learning curve for users new to distributed training
- Performance can be affected by network conditions in distributed setups
- Limited support for some advanced features in certain frameworks
Code Examples
- Basic Horovod initialization with TensorFlow:
import tensorflow as tf
import horovod.tensorflow as hvd
hvd.init()
# Pin GPU to be used to process local rank (one GPU per process)
gpus = tf.config.experimental.list_physical_devices('GPU')
for gpu in gpus:
tf.config.experimental.set_memory_growth(gpu, True)
if gpus:
tf.config.experimental.set_visible_devices(gpus[hvd.local_rank()], 'GPU')
- Distributed training with Keras:
import tensorflow as tf
import horovod.tensorflow.keras as hvd
hvd.init()
model = tf.keras.Sequential([...])
optimizer = tf.keras.optimizers.Adam(lr=0.001 * hvd.size())
optimizer = hvd.DistributedOptimizer(optimizer)
model.compile(optimizer=optimizer, loss='mse')
callbacks = [
hvd.callbacks.BroadcastGlobalVariablesCallback(0),
hvd.callbacks.MetricAverageCallback(),
]
model.fit(x_train, y_train, epochs=100, callbacks=callbacks)
- Distributed training with PyTorch:
import torch
import horovod.torch as hvd
hvd.init()
torch.cuda.set_device(hvd.local_rank())
model = torch.nn.Sequential(...)
optimizer = optim.SGD(model.parameters(), lr=0.01 * hvd.size())
hvd.broadcast_parameters(model.state_dict(), root_rank=0)
hvd.broadcast_optimizer_state(optimizer, root_rank=0)
optimizer = hvd.DistributedOptimizer(optimizer, named_parameters=model.named_parameters())
for epoch in range(100):
for batch_idx, (data, target) in enumerate(train_loader):
optimizer.zero_grad()
output = model(data)
loss = F.nll_loss(output, target)
loss.backward()
optimizer.step()
Getting Started
To get started with Horovod:
- Install Horovod:
pip install horovod
- Initialize Horovod in your code:
import horovod.tensorflow as hvd
hvd.init()
-
Modify your training script to use Horovod's distributed optimizers and callbacks.
-
Run your distributed training job:
horovodrun -np 4 python train.py
This command will run the training script on 4 processes, which can be on the same machine or distributed across multiple nodes.
Competitor Comparisons
An Open Source Machine Learning Framework for Everyone
Pros of TensorFlow
- Comprehensive ecosystem with tools for various ML tasks
- Extensive documentation and community support
- Flexible deployment options (mobile, web, cloud)
Cons of TensorFlow
- Steeper learning curve for beginners
- Can be slower for certain operations compared to Horovod
- More complex setup for distributed training
Code Comparison
TensorFlow (basic model):
import tensorflow as tf
model = tf.keras.Sequential([
tf.keras.layers.Dense(64, activation='relu'),
tf.keras.layers.Dense(10, activation='softmax')
])
model.compile(optimizer='adam', loss='categorical_crossentropy')
Horovod with TensorFlow:
import tensorflow as tf
import horovod.tensorflow as hvd
hvd.init()
model = tf.keras.Sequential([
tf.keras.layers.Dense(64, activation='relu'),
tf.keras.layers.Dense(10, activation='softmax')
])
opt = tf.optimizers.Adam(lr=0.001 * hvd.size())
opt = hvd.DistributedOptimizer(opt)
model.compile(optimizer=opt, loss='categorical_crossentropy')
The main difference is that Horovod adds distributed training capabilities to TensorFlow, allowing for easier scaling across multiple GPUs or machines. TensorFlow provides a more comprehensive framework, while Horovod focuses on optimizing distributed deep learning.
Tensors and Dynamic neural networks in Python with strong GPU acceleration
Pros of PyTorch
- More comprehensive deep learning framework with a wider range of features and tools
- Larger community and ecosystem, leading to better support and more resources
- Dynamic computational graph, allowing for more flexible and intuitive model development
Cons of PyTorch
- Steeper learning curve for beginners compared to Horovod's simplicity
- Can be slower for distributed training without additional optimizations
Code Comparison
PyTorch:
import torch
model = torch.nn.Linear(10, 1)
optimizer = torch.optim.SGD(model.parameters(), lr=0.01)
loss_fn = torch.nn.MSELoss()
for epoch in range(100):
optimizer.zero_grad()
output = model(input)
loss = loss_fn(output, target)
loss.backward()
optimizer.step()
Horovod:
import horovod.torch as hvd
hvd.init()
torch.cuda.set_device(hvd.local_rank())
optimizer = hvd.DistributedOptimizer(optimizer)
hvd.broadcast_parameters(model.state_dict(), root_rank=0)
for epoch in range(100):
optimizer.zero_grad()
output = model(input)
loss = loss_fn(output, target)
loss.backward()
optimizer.step()
Scalable, Portable and Distributed Gradient Boosting (GBDT, GBRT or GBM) Library, for Python, R, Java, Scala, C++ and more. Runs on single machine, Hadoop, Spark, Dask, Flink and DataFlow
Pros of XGBoost
- Highly efficient and scalable implementation of gradient boosting
- Supports a wide range of machine learning tasks (classification, regression, ranking)
- Built-in support for handling missing values and categorical features
Cons of XGBoost
- Less focused on distributed deep learning compared to Horovod
- May require more manual tuning of hyperparameters for optimal performance
- Limited support for neural network architectures
Code Comparison
XGBoost:
import xgboost as xgb
model = xgb.XGBClassifier()
model.fit(X_train, y_train)
predictions = model.predict(X_test)
Horovod:
import horovod.keras as hvd
hvd.init()
model = create_keras_model()
opt = keras.optimizers.Adam(lr=0.001 * hvd.size())
opt = hvd.DistributedOptimizer(opt)
model.compile(optimizer=opt, ...)
XGBoost focuses on gradient boosting algorithms, while Horovod is designed for distributed deep learning. XGBoost provides a simpler API for traditional machine learning tasks, whereas Horovod integrates with existing deep learning frameworks to enable distributed training. The choice between the two depends on the specific machine learning task and scalability requirements of the project.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Pros of DeepSpeed
- More comprehensive optimization techniques, including ZeRO (Zero Redundancy Optimizer)
- Better integration with PyTorch and Transformer models
- Supports a wider range of model parallelism techniques
Cons of DeepSpeed
- Steeper learning curve due to more complex features
- Less support for TensorFlow compared to Horovod
- May require more configuration for optimal performance
Code Comparison
Horovod:
import horovod.torch as hvd
hvd.init()
optimizer = optim.SGD(model.parameters())
optimizer = hvd.DistributedOptimizer(optimizer)
hvd.broadcast_parameters(model.state_dict(), root_rank=0)
DeepSpeed:
import deepspeed
model_engine, optimizer, _, _ = deepspeed.initialize(
args=args, model=model, model_parameters=params)
output = model_engine(input)
model_engine.backward(output)
model_engine.step()
Both libraries aim to improve distributed training performance, but DeepSpeed offers more advanced optimization techniques and better integration with PyTorch, while Horovod provides simpler implementation and broader framework support.
Open source platform for the machine learning lifecycle
Pros of MLflow
- Broader scope: Covers experiment tracking, model management, and deployment
- Language-agnostic: Supports multiple programming languages and frameworks
- User-friendly UI: Provides a web interface for easy experiment tracking and comparison
Cons of MLflow
- Less specialized for distributed training compared to Horovod
- May have higher overhead for simple use cases
- Requires additional setup and infrastructure for full functionality
Code Comparison
MLflow example:
import mlflow
with mlflow.start_run():
mlflow.log_param("learning_rate", 0.01)
mlflow.log_metric("accuracy", 0.85)
mlflow.sklearn.log_model(model, "model")
Horovod example:
import horovod.tensorflow as hvd
hvd.init()
optimizer = tf.optimizers.Adam(0.001 * hvd.size())
optimizer = hvd.DistributedOptimizer(optimizer)
with tf.GradientTape() as tape:
loss = compute_loss(model, x, y)
MLflow focuses on experiment tracking and model management, while Horovod specializes in distributed training. MLflow's code example shows logging parameters, metrics, and models, whereas Horovod's example demonstrates distributed optimization setup for TensorFlow.
Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
Pros of Ray
- More versatile, supporting distributed computing beyond just machine learning
- Offers a wider range of tools and libraries for various tasks (e.g., Ray Serve, Ray Tune)
- Easier to scale and manage complex distributed applications
Cons of Ray
- Steeper learning curve due to its broader scope
- May have more overhead for simple distributed machine learning tasks
- Less specialized for deep learning compared to Horovod
Code Comparison
Ray example:
import ray
@ray.remote
def f(x):
return x * x
futures = [f.remote(i) for i in range(4)]
print(ray.get(futures))
Horovod example:
import horovod.torch as hvd
hvd.init()
torch.cuda.set_device(hvd.local_rank())
optimizer = hvd.DistributedOptimizer(optimizer)
hvd.broadcast_parameters(model.state_dict(), root_rank=0)
Ray is more general-purpose, allowing for various distributed computing tasks, while Horovod is specifically designed for distributed deep learning. Ray's code focuses on task parallelism, whereas Horovod's code emphasizes data parallelism in deep learning frameworks.
Convert designs to code with AI
Introducing Visual Copilot: A new AI model to turn Figma designs to high quality code using your components.
Try Visual CopilotREADME
.. raw:: html
<p align="center"><img src="https://user-images.githubusercontent.com/16640218/34506318-84d0c06c-efe0-11e7-8831-0425772ed8f2.png" alt="Logo" width="200"/></p>
<br/>
Horovod
.. raw:: html
.. image:: https://badge.fury.io/py/horovod.svg :target: https://badge.fury.io/py/horovod :alt: PyPI Version
.. image:: https://badge.buildkite.com/6f976bc161c69d9960fc00de01b69deb6199b25680a09e5e26.svg?branch=master :target: https://buildkite.com/horovod/horovod :alt: Build Status
.. image:: https://readthedocs.org/projects/horovod/badge/?version=latest :target: https://horovod.readthedocs.io/en/latest/ :alt: Documentation Status
.. image:: https://img.shields.io/badge/slack-chat-green.svg?logo=slack :target: https://forms.gle/cPGvty5hp31tGfg79 :alt: Slack
.. raw:: html
.. raw:: html
.. image:: https://img.shields.io/badge/License-Apache%202.0-blue.svg :target: https://img.shields.io/badge/License-Apache%202.0-blue.svg :alt: License
.. image:: https://app.fossa.com/api/projects/git%2Bgithub.com%2Fhorovod%2Fhorovod.svg?type=shield :target: https://app.fossa.com/projects/git%2Bgithub.com%2Fhorovod%2Fhorovod?ref=badge_shield :alt: FOSSA Status
.. image:: https://bestpractices.coreinfrastructure.org/projects/2373/badge :target: https://bestpractices.coreinfrastructure.org/projects/2373 :alt: CII Best Practices
.. image:: https://pepy.tech/badge/horovod :target: https://pepy.tech/project/horovod :alt: Downloads
.. raw:: html
.. inclusion-marker-start-do-not-remove
|
Horovod is a distributed deep learning training framework for TensorFlow, Keras, PyTorch, and Apache MXNet. The goal of Horovod is to make distributed deep learning fast and easy to use.
.. raw:: html
Horovod is hosted by the LF AI & Data Foundation <https://lfdl.io>
_ (LF AI & Data). If you are a company that is deeply
committed to using open source technologies in artificial intelligence, machine, and deep learning, and want to support
the communities of open source projects in these domains, consider joining the LF AI & Data Foundation. For details
about who's involved and how Horovod plays a role, read the Linux Foundation announcement <https://lfdl.io/press/2018/12/13/lf-deep-learning-welcomes-horovod-distributed-training-framework-as-newest-project/>
_.
|
.. contents::
|
Documentation
Latest Release <https://horovod.readthedocs.io/en/stable>
_master <https://horovod.readthedocs.io/en/latest>
_
|
Why Horovod?
The primary motivation for this project is to make it easy to take a single-GPU training script and successfully scale it to train across many GPUs in parallel. This has two aspects:
- How much modification does one have to make to a program to make it distributed, and how easy is it to run it?
- How much faster would it run in distributed mode?
Internally at Uber we found the MPI model to be much more straightforward and require far less code changes than previous
solutions such as Distributed TensorFlow with parameter servers. Once a training script has been written for scale with
Horovod, it can run on a single-GPU, multiple-GPUs, or even multiple hosts without any further code changes.
See the Usage <#usage>
__ section for more details.
In addition to being easy to use, Horovod is fast. Below is a chart representing the benchmark that was done on 128 servers with 4 Pascal GPUs each connected by RoCE-capable 25 Gbit/s network:
.. image:: https://user-images.githubusercontent.com/16640218/38965607-bf5c46ca-4332-11e8-895a-b9c137e86013.png :alt: 512-GPU Benchmark
Horovod achieves 90% scaling efficiency for both Inception V3 and ResNet-101, and 68% scaling efficiency for VGG-16.
See Benchmarks <docs/benchmarks.rst>
_ to find out how to reproduce these numbers.
While installing MPI and NCCL itself may seem like an extra hassle, it only needs to be done once by the team dealing with infrastructure, while everyone else in the company who builds the models can enjoy the simplicity of training them at scale.
Install
To install Horovod on Linux or macOS:
- Install
CMake <https://cmake.org/install/>
__
.. raw:: html
<p/>
2. If you've installed TensorFlow from PyPI <https://pypi.org/project/tensorflow>
__, make sure that g++-5
or above is installed.
Starting with TensorFlow 2.10 a C++17-compliant compiler like g++8
or above will be required.
If you've installed PyTorch from PyPI <https://pypi.org/project/torch>
__, make sure that g++-5
or above is installed.
If you've installed either package from Conda <https://conda.io>
_, make sure that the gxx_linux-64
Conda package is installed.
.. raw:: html
<p/>
3. Install the horovod
pip package.
To run on CPUs:
.. code-block:: bash
$ pip install horovod
To run on GPUs with NCCL:
.. code-block:: bash
$ HOROVOD_GPU_OPERATIONS=NCCL pip install horovod
For more details on installing Horovod with GPU support, read Horovod on GPU <docs/gpus.rst>
_.
For the full list of Horovod installation options, read the Installation Guide <docs/install.rst>
_.
If you want to use MPI, read Horovod with MPI <docs/mpi.rst>
_.
If you want to use Conda, read Building a Conda environment with GPU support for Horovod <docs/conda.rst>
_.
If you want to use Docker, read Horovod in Docker <docs/docker.rst>
_.
To compile Horovod from source, follow the instructions in the Contributor Guide <docs/contributors.rst>
_.
Concepts
Horovod core principles are based on MPI <http://mpi-forum.org/>
_ concepts such as size, rank,
local rank, allreduce, allgather, broadcast, and alltoall. See this page <docs/concepts.rst>
_
for more details.
Supported frameworks
See these pages for Horovod examples and best practices:
Horovod with TensorFlow <docs/tensorflow.rst>
_Horovod with XLA in Tensorflow <xla.rst>
_Horovod with Keras <docs/keras.rst>
_Horovod with PyTorch <docs/pytorch.rst>
_Horovod with MXNet <docs/mxnet.rst>
_
Usage
To use Horovod, make the following additions to your program:
- Run
hvd.init()
to initialize Horovod.
.. raw:: html
<p/>
2. Pin each GPU to a single process to avoid resource contention.
With the typical setup of one GPU per process, set this to local rank. The first process on the server will be allocated the first GPU, the second process will be allocated the second GPU, and so forth.
.. raw:: html
<p/>
3. Scale the learning rate by the number of workers.
Effective batch size in synchronous distributed training is scaled by the number of workers. An increase in learning rate compensates for the increased batch size.
.. raw:: html
<p/>
4. Wrap the optimizer in hvd.DistributedOptimizer
.
The distributed optimizer delegates gradient computation to the original optimizer, averages gradients using allreduce or allgather, and then applies those averaged gradients.
.. raw:: html
<p/>
5. Broadcast the initial variable states from rank 0 to all other processes.
This is necessary to ensure consistent initialization of all workers when training is started with random weights or restored from a checkpoint.
.. raw:: html
<p/>
6. Modify your code to save checkpoints only on worker 0 to prevent other workers from corrupting them.
.. raw:: html
<p/>
Example using TensorFlow v1 (see the examples <https://github.com/horovod/horovod/blob/master/examples/>
_ directory for full training examples):
.. code-block:: python
import tensorflow as tf
import horovod.tensorflow as hvd
# Initialize Horovod
hvd.init()
# Pin GPU to be used to process local rank (one GPU per process)
config = tf.ConfigProto()
config.gpu_options.visible_device_list = str(hvd.local_rank())
# Build model...
loss = ...
opt = tf.train.AdagradOptimizer(0.01 * hvd.size())
# Add Horovod Distributed Optimizer
opt = hvd.DistributedOptimizer(opt)
# Add hook to broadcast variables from rank 0 to all other processes during
# initialization.
hooks = [hvd.BroadcastGlobalVariablesHook(0)]
# Make training operation
train_op = opt.minimize(loss)
# Save checkpoints only on worker 0 to prevent other workers from corrupting them.
checkpoint_dir = '/tmp/train_logs' if hvd.rank() == 0 else None
# The MonitoredTrainingSession takes care of session initialization,
# restoring from a checkpoint, saving to a checkpoint, and closing when done
# or an error occurs.
with tf.train.MonitoredTrainingSession(checkpoint_dir=checkpoint_dir,
config=config,
hooks=hooks) as mon_sess:
while not mon_sess.should_stop():
# Perform synchronous training.
mon_sess.run(train_op)
Running Horovod
The example commands below show how to run distributed training.
See Run Horovod <docs/running.rst>
_ for more details, including RoCE/InfiniBand tweaks and tips for dealing with hangs.
-
To run on a machine with 4 GPUs:
.. code-block:: bash
$ horovodrun -np 4 -H localhost:4 python train.py
-
To run on 4 machines with 4 GPUs each:
.. code-block:: bash
$ horovodrun -np 16 -H server1:4,server2:4,server3:4,server4:4 python train.py
-
To run using Open MPI without the
horovodrun
wrapper, seeRunning Horovod with Open MPI <docs/mpi.rst>
_. -
To run in Docker, see
Horovod in Docker <docs/docker.rst>
_. -
To run on Kubernetes, see
Helm Chart <https://github.com/horovod/horovod/tree/master/docker/helm/>
,Kubeflow MPI Operator <https://github.com/kubeflow/mpi-operator/>
,FfDL <https://github.com/IBM/FfDL/tree/master/etc/examples/horovod/>
, andPolyaxon <https://docs.polyaxon.com/integrations/horovod/>
. -
To run on Spark, see
Horovod on Spark <docs/spark.rst>
_. -
To run on Ray, see
Horovod on Ray <docs/ray.rst>
_. -
To run in Singularity, see
Singularity <https://github.com/sylabs/examples/tree/master/machinelearning/horovod>
_. -
To run in a LSF HPC cluster (e.g. Summit), see
LSF <docs/lsf.rst>
_. -
To run on Hadoop Yarn, see
TonY <https://github.com/linkedin/TonY/>
_.
Gloo
Gloo <https://github.com/facebookincubator/gloo>
_ is an open source collective communications library developed by Facebook.
Gloo comes included with Horovod, and allows users to run Horovod without requiring MPI to be installed.
For environments that have support both MPI and Gloo, you can choose to use Gloo at runtime by passing the --gloo
argument to horovodrun
:
.. code-block:: bash
$ horovodrun --gloo -np 2 python train.py
mpi4py
Horovod supports mixing and matching Horovod collectives with other MPI libraries, such as mpi4py <https://mpi4py.scipy.org>
_,
provided that the MPI was built with multi-threading support.
You can check for MPI multi-threading support by querying the hvd.mpi_threads_supported()
function.
.. code-block:: python
import horovod.tensorflow as hvd
# Initialize Horovod
hvd.init()
# Verify that MPI multi-threading is supported.
assert hvd.mpi_threads_supported()
from mpi4py import MPI
assert hvd.size() == MPI.COMM_WORLD.Get_size()
You can also initialize Horovod with an mpi4py
sub-communicator, in which case each sub-communicator
will run an independent Horovod training.
.. code-block:: python
from mpi4py import MPI
import horovod.tensorflow as hvd
# Split COMM_WORLD into subcommunicators
subcomm = MPI.COMM_WORLD.Split(color=MPI.COMM_WORLD.rank % 2,
key=MPI.COMM_WORLD.rank)
# Initialize Horovod
hvd.init(comm=subcomm)
print('COMM_WORLD rank: %d, Horovod rank: %d' % (MPI.COMM_WORLD.rank, hvd.rank()))
Inference
Learn how to optimize your model for inference and remove Horovod operations from the graph here <docs/inference.rst>
_.
Tensor Fusion
One of the unique things about Horovod is its ability to interleave communication and computation coupled with the ability to batch small allreduce operations, which results in improved performance. We call this batching feature Tensor Fusion.
See here <docs/tensor-fusion.rst>
__ for full details and tweaking instructions.
Horovod Timeline
Horovod has the ability to record the timeline of its activity, called Horovod Timeline.
.. image:: https://user-images.githubusercontent.com/16640218/29735271-9e148da0-89ac-11e7-9ae0-11d7a099ac89.png :alt: Horovod Timeline
Use Horovod timeline to analyze Horovod performance.
See here <docs/timeline.rst>
__ for full details and usage instructions.
Automated Performance Tuning
Selecting the right values to efficiently make use of Tensor Fusion and other advanced Horovod features can involve
a good amount of trial and error. We provide a system to automate this performance optimization process called
autotuning, which you can enable with a single command line argument to horovodrun
.
See here <docs/autotune.rst>
__ for full details and usage instructions.
Horovod Process Sets
Horovod allows you to concurrently run distinct collective operations in different groups of processes taking part in
one distributed training. Set up hvd.process_set
objects to make use of this capability.
See Process Sets <docs/process_set.rst>
__ for detailed instructions.
Guides
- Run distributed training in Microsoft Azure using
Batch AI and Horovod <https://github.com/Azure/BatchAI/tree/master/recipes/Horovod>
_. Distributed model training using Horovod <https://spell.ml/blog/distributed-model-training-using-horovod-XvqEGRUAACgAa5th>
_.
Send us links to any user guides you want to publish on this site
Troubleshooting
See Troubleshooting <docs/troubleshooting.rst>
_ and submit a ticket <https://github.com/horovod/horovod/issues/new>
_
if you can't find an answer.
Citation
Please cite Horovod in your publications if it helps your research:
::
@article{sergeev2018horovod,
Author = {Alexander Sergeev and Mike Del Balso},
Journal = {arXiv preprint arXiv:1802.05799},
Title = {Horovod: fast and easy distributed deep learning in {TensorFlow}},
Year = {2018}
}
Publications
-
Sergeev, A., Del Balso, M. (2017) Meet Horovod: Uberâs Open Source Distributed Deep Learning Framework for TensorFlow. Retrieved from
https://eng.uber.com/horovod/ <https://eng.uber.com/horovod/>
_ -
Sergeev, A. (2017) Horovod - Distributed TensorFlow Made Easy. Retrieved from
https://www.slideshare.net/AlexanderSergeev4/horovod-distributed-tensorflow-made-easy <https://www.slideshare.net/AlexanderSergeev4/horovod-distributed-tensorflow-made-easy>
_ -
Sergeev, A., Del Balso, M. (2018) Horovod: fast and easy distributed deep learning in TensorFlow. Retrieved from
arXiv:1802.05799 <https://arxiv.org/abs/1802.05799>
_
References
The Horovod source code was based off the Baidu tensorflow-allreduce <https://github.com/baidu-research/tensorflow-allreduce>
_
repository written by Andrew Gibiansky and Joel Hestness. Their original work is described in the article
Bringing HPC Techniques to Deep Learning <http://andrew.gibiansky.com/blog/machine-learning/baidu-allreduce/>
_.
Getting Involved
Community Slack <https://forms.gle/cPGvty5hp31tGfg79>
_ for collaboration and discussionHorovod Announce <https://lists.lfai.foundation/g/horovod-announce>
_ for updates on the projectHorovod Technical-Discuss <https://lists.lfai.foundation/g/horovod-technical-discuss>
_ for public discussionHorovod Security <https://lists.lfai.foundation/g/horovod-security>
_ to report security vulnerabilities
.. inclusion-marker-end-do-not-remove Place contents above here if they should also appear in read-the-docs. Contents below are already part of the read-the-docs table of contents.
Top Related Projects
An Open Source Machine Learning Framework for Everyone
Tensors and Dynamic neural networks in Python with strong GPU acceleration
Scalable, Portable and Distributed Gradient Boosting (GBDT, GBRT or GBM) Library, for Python, R, Java, Scala, C++ and more. Runs on single machine, Hadoop, Spark, Dask, Flink and DataFlow
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Open source platform for the machine learning lifecycle
Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
Convert designs to code with AI
Introducing Visual Copilot: A new AI model to turn Figma designs to high quality code using your components.
Try Visual Copilot