Skip to content

Latest commit

 

History

History
7 lines (5 loc) · 378 Bytes

README.md

File metadata and controls

7 lines (5 loc) · 378 Bytes

Distributed-TPU-Training

Google Colab offers options to use GPUs and TPUs as hardware accelerators.
While this can already significantly increase the training speed, we can increase the speed even further by distributing trainig over a pool of these accelerators.

This is an illustration of how to define a distribution strategy and train a model in a distributed way.