Training delivery in the cloud: cloud machine learning engine In the previous episode, we talked about the problems we face when your dataset is too big to fit on your local machine, and we discussed how we can move data to the cloud, with scalable storage. Today we are in the second half of that problem - those computer resources are falling apart. When training large models, the current approach involves training in parallel. Our data is split and sent to multiple working machines, and then the model must keep the information and signals that it is receiving from each machine, again together, to create a fully trained model. Do you like configuration If you wish, you can configure them to spin some virtual machines, install the necessary libraries, network them together, and run distributed machine learning. And then when you're done, you want to be sure to take those machines down. While some may find it easy on the surface, it can be challenging if you are not familiar with thi...