Articles
Deployment of a TensorFlow model to Kubernetes
Let’s imagine that you’ve just finished training your new TensorFlow model and want to start using it in your application(s). One obvious way to do so is to simply import it in the source code of every application that uses it. However, it might be more versatile to keep your model in one place as standalone and simply have applications exchange data with it through API calls. This article will go through the steps of building such a system and deploy the result to Kubernetes.
Tensorflow 1.X basics
Here are the basics steps necessary to write TensorFlow code.
Gitlab CI commands for TF serving
This is an example .gitlab-ci.yml file which can be used to containerize and deploy a tensorflow model
Serving a Keras model using Tensorflow serving and Docker
A Keras model can be created in various ways, for example using the <a href="https://keras.io/getting-started/sequential-model-guide/" rel="noopener noreferrer nofollow">sequential model</a>:
0 - 10 / 4