In this chapter, we will learn how to deploy the trained model in the various platforms for maximum throughput and minimum latency. We will understand performance on various hardware such as a GPU and a CPU. We will follow the steps of deploying TensorFlow in platforms such as Amazon Web Services, Google Cloud Platform, and mobile platforms such as Android, iOS, and Tegra.
We will cover the following topics in this chapter:
- Understanding the factors affecting the performance of the deep learning model training and inference
- Improving the performance through various methods
- Seeing the benchmarks of various hardware and learning the steps to tweak them for maximum performance
- Using various cloud platforms for deployment
- Using various mobile platforms for deployment