Chainer
6.0.0b2
Tutorials
Chainer at a Glance
Concepts Walkthrough
Examples
Neural Net Examples
Colab Notebook Examples
Awesome Chainer
References
API Reference
Installation
Distributed Deep Learning with ChainerMN
Installation
Tutorial
Overview
Step 1: Communicators and Optimizers
Step 2: Datasets and Evaluators
Tips and FAQs
Model Parallel
API Reference
Other
API Compatibility Policy
Contribution Guide
Tips and FAQs
Performance Best Practices
Upgrade Guide
Comparison with Other Frameworks
License
Community
Slack Chat
Forums
Chainer
Docs
»
Distributed Deep Learning with ChainerMN
»
Tutorial
Tutorial
¶
Overview
Data Parallelism
Training Iterations
MPI
Step 1: Communicators and Optimizers
Creating a Communicator
Creating a Multi-Node Optimizer
Run
Multi-node execution
Copying datasets
If you have trouble
Next Steps
Step 2: Datasets and Evaluators
Scattering Datasets
Creating A Multi-Node Evaluator
Suppressing Unnecessary Extensions
Tips and FAQs
Using MultiprocessIterator
Using Your Own Evaluator
Using MPI4py Communicator
Using FP16
MPI process hangs after an unhandled Python exception.