Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.
Note
titleUnder construction

This page is currently under detailed update as our previous cluster Topaz has beendecommissioned and the content of the page needs to be updated for its use on Setonix.

In this tutorial, you are going to see how to write a Horovod-powered distributed TensorFlow computation. More specifically, the final goal is to train different models in parallel by assigning each of them to a different GPU. The discussion is organised in two sections. The first section illustrates Horovod's basic concepts and its usage coupled with TensorFlow, the second one uses the MNIST classification task as test case.

...