Child pages
  • HAL Benchmarks 2020
Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 3 Next »

ImageNet Distributed Mixed-precision Training with PyTorch

References

Source code: https://github.com/richardkxu/distributed-pytorch

Overview

We will cover the following training methods for PyTorch:

  • regular, single node, single GPU training
  • torch.nn.DataParallel
  • torch.nn.DistributedDataParallel
  • mixed precision training with NVIDIA Apex
  • TensorBoard logging under distributed training context

We will cover the following use cases:

  • Single node single GPU training
  • Single node multi-GPU training
  • Multi-node multi-GPU training




  • No labels