paint-brush
Lessons for Improving Training Performance — Part 1by@emwatz
977 reads
977 reads

Lessons for Improving Training Performance — Part 1

by Emily Watkins7mOctober 2nd, 2019
Read on Terminal Reader
Read this story w/o Javascript
tldt arrow

Too Long; Didn't Read

Pure Storage published TensorFlow deep learning performance results in March. In Part 2 we’ll investigate how input pipelines affect overall training throughput. Performance gains came from ten months of application developments, not a single factor. With FP16 support, developers can take advantage of Tensor Cores present on Nvidia GPUs, trading lower precision for higher training throughput. With larger batch sizes, more samples are processed together, amortizing coordination work. The input pipeline during training, previously a performance limiter, is more efficient.

Companies Mentioned

Mention Thumbnail
Mention Thumbnail
featured image - Lessons for Improving Training Performance — Part 1
Emily Watkins HackerNoon profile picture
Emily Watkins

Emily Watkins

@emwatz

L O A D I N G
. . . comments & more!

About Author

Emily Watkins HackerNoon profile picture
Emily Watkins@emwatz

TOPICS

THIS ARTICLE WAS FEATURED IN...

Permanent on Arweave
Read on Terminal Reader
Read this story in a terminal
 Terminal
Read this story w/o Javascript
Read this story w/o Javascript
 Lite
Campaign-archive
Tefter
Tefter
Faq-courses
Garker