Toby Chan
Toby Chan
Feb 24, 2017 · 1 min read

Great post and thanks for the contribution of serving multiple model versions.

For scalability and fail over, I would like to advice using Kubernetes. You can build the docker image of TensorFlow Serving and just download the one from Google. With Kubernetes, you can deploy the models with one command and leverage the advantages of load balance and fault tolerance.

Actually, we have implement the Cloud Machine Learning service, just like Google CloudML. And everyone can submit the train job or create the model service very easily.

    Toby Chan

    Written by

    Toby Chan

    Welcome to a place where words matter. On Medium, smart voices and original ideas take center stage - with no ads in sight.
    Follow all the topics you care about, and we’ll deliver the best stories for you to your homepage and inbox.
    Get unlimited access to the best stories on Medium — and support writers while you’re at it. Just $5/month.