Articles

Docker restart container when docker restarts

Simply add the following flag when using docker run
TutorialsDocker

Docker images and containers management

Here are a few commands to manage docker images and containers
DockerTutorials

Docker behind a proxy

Docker does not use environment variables for proxy configuration. This article presents how to configure Docker to use a proxy.
DockerTutorials

Docker HTTP (insecure) registry

By default Docker refuses to push/pull from registries that are not served using HTTPS.
DevOpsDockerTutorials

Divide container in equally sized divs

Let's imagine a container with three divs as content:
CSSTutorials

Distributing a Helm chart on Artifact Hub

Distributing a Helm chart on Artifact Hub thumbnail
Building applications in a microservice architecture has become more and more popular recently. With this design pattern, an application is composed of multiple services that run independently and generally share data across network protocols.
KubernetesHelmWIPDevOps

Dissecting a Kubernetes manifest

Dissecting a Kubernetes manifest thumbnail
Kubernetes manifests can seem quite daunting at first, but it is important to understand that their apparent complexity is simply a result of the large number of customization options. In the end, manifests are used to deploy resources that interact with each other, which, among others, lead to the correct operations of containerized applications. Consequently, resources specified in a manifest must be configured accordingly. This article aims at explaining how manifests are structured to do so.
TutorialsKubernetesFeatured

Deployment of a TensorFlow model to Kubernetes

Let’s imagine that you’ve just finished training your new TensorFlow model and want to start using it in your application(s). One obvious way to do so is to simply import it in the source code of every application that uses it. However, it might be more versatile to keep your model in one place as standalone and simply have applications exchange data with it through API calls. This article will go through the steps of building such a system and deploy the result to Kubernetes.
AI / MLDevOpsTensorFlow

Deploying a TensorFlow model on a Jetson Nano using TensorFlow serving and K3s

The Nvidia Jetson Nano constitutes a low cost platform for AI applications, ideal for edge computing.However, due to the architecture of its CPU, deploying applications to the SBC can be challenging. In this guide, we'll install and configure K3s, a lightweight kubernetes distribution made specifically for edge devices. Once done we'll build and deploy an TensorFlow model in the K3s cluster.
K3sTutorialsKubernetes

Deploy a Neo4J instance in Kubernetes

Using this manifest, a Neo4J instance can be deployed in a Kubernetes cluster
KubernetesNeo4J