Description of the eHealth development and deployment cycle, including description of the use of eHealth environments and flow.
Overview of build and deployment
Applications must be deployed using one of the predefined eHealth Helm chart and deployed in eHealth Service Mesh
Telemedicine Solutions provider must use the eHealth Infrastructure Deployment Pipeline for deploying to the eHealth environments.
- Telemedicine Solutions provider shall use the official eHealth docker images and helm chart for application definition and development
- After local development, the Telemedicine Solutions Providers shall upload the docker images to Docker and Helm chart repository (Harbor)
- Telemedicine Solutions Providers uses GitLab to define the deployment file or "desired state file" based on Helmsman
- A configured Jenkins deployment pipeline monitors the Helmans file in GitLab and push updates to Kubernetes
- The Jenkins deployment pipeline is configured by the eHealth Infrastructure provider.
- Kubernetes control loop watch the state of your cluster, then make or request changes where needed
The following figure illustrates the deployment pipeline.
Development and deployment happens in four phases.
Phase #1 Local development
All Telemedicine Solution Providers shall start development of their components in their own local environment.
This could be a complete clone of the official test environments running on the eHealth-platform. But this is not a strict requirement.
Phase #2 Publish build to the eHealth-platform
When the Solution Providers supplier believes that the application / service / microservice is ready for testing it is published to the eHealth-platform.
This means that the docker image is signed and pushed to the central docker image registry used and hosted by the eHealth-platform.
Phase #3 Deploy to test environment
To actually deploy the docker image as a container on the first test environment, the image needs to be added to the helmsman specification file for the given environment.
All applications running on an environment is specified as code in the desired state specification for the given environment.
For each application this includes:
- The docker image to run
- Docker repository
- Image tag
- The helm chart
- Helm chart repository
- Helm chart version
- Configuration
- Ports
- Replicas
- Memory usage
- Environment variables
- Database and queue secrets
- DNS bindings
- etc.
When the desired state specification is updated the applications and configuration is automatically rolled on to the environment.
Phase #4 Test and Promotion
Notice, every release must follow a strict release plan where eHealth environments are visited in the required order, that is deployed and tested on EXTTEST, before being promoted to PREPRODUCTION, and finally deployed to PRODUCTION.
External Test
Tests should be carried out on the Test environment.
When the application has passed QA, it can be promoted to the next environment. This happens by a promotion of a specific desired state specification in Jenkins-test, by users with the right privileges.
PreProd and Prod
Going onwards to preprod
happens by making a pull-request to the "prod
" branch from the "master
" branch.
When the pull-request is approved all updates in the desired state specification for preprod
is rolled on the preprod
environment.
Deployment to production happens by a promotion of a specific desired state specification in Jenkins-prod, by users with the right privileges.
Hot-fixing
If serious bugs or security errors is found in production code, or containers a hotfix can be handled in the following way.
Rollback
Rollback is handled somewhat semi-automatic, if the problem is visible through the pod probe services liveness and readiness. In other rollback situation manual intervention is necessary.
In the first situation, where either liveness and readiness is responding negative, the new pod instance will never be part of the serving cluster. The cluster will automatically keep sending traffic to the old instances of the service and skip the new pod. What should happen next has to be decided by the one who started the deployment. It could be a configuration fix or a rollback to the old software version.
In the second situation, where both liveness and readiness respond positive, the new pod will receive trafik. The one who started the deployment will look at cluster readout and decide a rollback. That person has to commit a new cluster configuration. This could be a reverse commit to the desired state file.
Note: if we are using Canary deployment, the negative impact of a software problem can be held at a minimum. That is if the negative impact can be found in prometheus.
Diagram of the flow
The flow can allows be viewed as follows.
- Application Vendor perform Local Development
- Application Vendor pust application and docker images to the FUT platform
- The eHealth infrastructure deploy to different environments, based on the Helmsman files provided by the application vendor.
- Application Vendor (or Customer) can test the Telemedicine solution on EXTTEST
- After successul tests, the build can then be promoted to next eHealth environment (PREPROCUTION).
- After successul tests, the build can then be promoted to eHealth PRODUCTION environment.
In the figure is is shown as SRE Team (Software Reliable Engineering) must approve deployment to external test, pre-prod and production. This is not the case for external applications, furthermore, external applications do not have access to Internal Test (INTTEST), but has a special environment called DEVENVCGI, for internal test.