Distributed TensorFlow training (Google I/O '18)

Loading

Follow to receive video recommendations   a   A


To efficiently train machine learning models, you will often need to scale your training to multiple GPUs, or even multiple machines. TensorFlow now offers rich functionality to achieve this with just a few lines of code. Join this session to learn how to set this up. Rate this session by signing-in on the I/O website here → https://goo.gl/sBZMEm Distribution Strategy API: https://goo.gl/F9vXqQ https://goo.gl/Zq2xvJ ResNet50 Model Garden example with MirroredStrategy API: https://goo.gl/3UWhj8 Performance Guides: https://goo.gl/doqGE7 https://goo.gl/NCnrCn Commands to set up a GCE instance and run distributed training: https://goo.gl/xzwN4C Multi-machine distributed training with train_and_evaluate: https://goo.gl/kyikAC Watch more TensorFlow sessions from I/O '18 here → https://goo.gl/GaAnBR See all the sessions from Google I/O '18 here → https://goo.gl/q1Tr8x Subscribe to the TensorFlow channel → https://goo.gl/ht3WGe #io18

For GoLang training courses, please contact Wojtek.