copyright | lastupdated | keywords | subcollection | ||
---|---|---|---|---|---|
|
2024-10-09 |
kubernetes, openshift, red hat, red hat openshift, update, upgrade, containers |
containers |
{{site.data.keyword.attribute-definition-list}}
{: #update_app}
Use Kubernetes-native and third-party tools to perform rolling updates and rollbacks of apps without downtime for your users. {: shortdesc}
{: #updating_apps}
To update your app, you can choose from various strategies such as the following. You might start with a rolling deployment or instantaneous switch before you progress to a more complicated canary deployment.
Rolling deployment
: You can use Kubernetes-native functionality to create a v2
deployment and to gradually replace your previous v1
deployment. This approach requires that apps are backwards-compatible so that users who are served the v2
app version don't experience any breaking changes. For more information, see Managing rolling deployments to update your apps.
Instantaneous switch
: Also referred to as a blue-green deployment, an instantaneous switch requires double the compute resources to have two versions of an app running at once. With this approach, you can switch your users to the newer version in near real time. Make sure that you use service label selectors (such as version: green
and version: blue
) to make sure that requests are sent to the correct app version. You can create the new version: green
deployment, wait until it is ready, and then delete the version: blue
deployment. Or you can perform a rolling update, but set the maxUnavailable
parameter to 0%
and the maxSurge
parameter to 100%
.
Canary or A/B deployment
: A more complex update strategy, a canary deployment is when you pick a percentage of users such as 5% and send them to the new app version. You collect metrics in your logging and monitoring tools on how the new app version performs, do A/B testing, and then roll out the update to more users. As with all deployments, labeling the app (such as version: stable
and version: canary
) is critical. To manage canary deployments, you might install the managed Istio add-on service mesh, set up {{site.data.keyword.mon_short}} for your cluster, and then use the Istio service mesh for A/B testing as described in this blog post{: external}.
{: #app_scaling}
With Kubernetes, you can enable horizontal pod autoscaling{: external} to automatically increase or decrease the number of instances of your apps based on CPU. {: shortdesc}
Want to scale your worker nodes instead of your pods? Check out the cluster autoscaler. {: tip}
Before you begin
- Log in to your account. If applicable, target the appropriate resource group. Set the context for your cluster.
- Make sure that you are assigned a service access role that grants the appropriate Kubernetes RBAC role so that you can work with Kubernetes resources in the namespace.
To scale your apps:
-
Deploy your app to a cluster from the CLI. For more complex deployments, you might need to create a configuration file.
kubectl create deployment <app_name> --image=<image>
{: pre}
-
Set CPU resource limits in millicores for the containers that run in your deployment, such as
100m
. You can also set memory limits, but the horizontal pod autoscaler only considers CPU resource limits for scaling purposes. For more information, see thekubectl set resources
documentation{: external}.kubectl set resources deployment <app_name> --limits=cpu=100m
{: pre}
-
Create a horizontal pod autoscaler and define your policy. For more information, see the
kubectl autoscale
documentation{: external}.kubectl autoscale deployment <deployment_name> --cpu-percent=<percentage> --min=<min_value> --max=<max_value>
{: pre}
Options Description --cpu-percent
The average CPU utilization that is maintained by the Horizontal Pod Autoscaler, which is specified as a percentage. --min
The minimum number of deployed pods that are used to maintain the specified CPU utilization percentage. --max
The maximum number of deployed pods that are used to maintain the specified CPU utilization percentage. {: caption="Understanding your command options" caption-side="bottom"}
{: #app_rolling}
You can manage the rollout of your app changes in an automated and controlled fashion for workloads with a pod template such as deployments. If your rollout isn't going according to plan, you can roll back your deployment to the previous revision. {: shortdesc}
Want to prevent downtime during your rolling update? Be sure to specify a readiness probe in your deployment so that the rollout proceeds to the next app pod after the most recently updated pod is ready. {: tip}
Before you begin
- Log in to your account. If applicable, target the appropriate resource group. Set the context for your cluster.
- Create a deployment.
- Make sure that you have a service access role that grants the appropriate Kubernetes RBAC role so that you can work with Kubernetes resources in the namespace.
To manage rolling updates to your apps:
-
To make sure that your deployments are marked as ready only when the container is running and ready to service requests, add liveness and readiness probes to your deployment.
-
Update your deployment to include a rolling update strategy that specifies the maximum surge and unavailable pods or percentage of pods during the update.
apiVersion: apps/v1 kind: Deployment metadata: name: nginx-test spec: replicas: 10 selector: matchLabels: service: http-server minReadySeconds: 5 progressDeadlineSeconds: 600 strategy: type: RollingUpdate rollingUpdate: maxUnavailable: 50% maxSurge: 2 ...
{: codeblock}
spec.minReadySeconds
: By default, deployments wait until the pod is marked asready
to continue with the rollout. If you notice that the deployment continues to create pods even though your app in the most recent pod is not yet ready, use this field to slow down the deployment rollout. For example, if you specify5
, the deployment waits for 5 seconds after the pod isready
before it creates the next pod.spec.progressDeadlineSeconds
: Set a timeout in seconds before a deployment is considered failed. For example, without a timeout, if your new app version has a bug and hangs immediately, the rollout can't continue because the pod never reaches aready
state. If you set this timeout to600
seconds, then if any phase of the rollout fails to progress for 10 minutes, the deployment is marked as failed and the rollout stops.spec.strategy.type
: Specify theRollingUpdate
strategy type.spec.strategy.rollingUpdate.maxUnavailable
: Set the maximum number of pods that can be unavailable during an update, as a number (2
) or percentage (50%
). Generally, use a percentage so that if you change the number of replicas later you don't have to remember to update the number here, unless you want to limit the rollout to allow only one pod to be down at a time. If you never want to fall under 100% capacity, set this value to0%
and specify thespec.strategy.type.rollingUpdate.maxSurge
parameter.spec.strategy.rollingUpdate.maxSurge
: Set how many extra resources the deployment can use during the rollout, as a number (2
) or percentage (50%
). For example, if your deployment specifies10
replicas and you set themaxSurge
to2
, then during the rollout, two new replicas are created. You now have 12 replicas (10 existing, 2 new). After the two new replicas are ready, the deployment scales down the old replicas to 8 to meet the specified 10 replicas. This process continues until the rollout is complete and all 10 replicas run the new version.If you want to perform a blue-green instantaneous switch style update, set the
maxSurge
to100%
. The deployment creates all the new required replicas, then scales down the old version replicas to 0. {: tip} -
Roll out{: external} a change. For example, you might want to change the image that you used in your initial deployment.
-
Get the deployment name.
kubectl get deployments
{: pre}
-
Get the pod name.
kubectl get pods
{: pre}
-
Get the name of the container that runs in the pod.
kubectl describe pod <pod_name>
{: pre}
-
Set the new image for the deployment to use.
kubectl set image deployment/<deployment_name><container_name>=<image_name>
{: pre}
When you run the commands, the change is immediately applied and logged in the roll-out history.
-
-
Check the status of your deployment.
kubectl rollout status deployments/<deployment_name>
{: pre}
If you notice something in the status that you want time to follow up on, you can pause and resume your rollout with the following commands.
kubectl rollout pause deployment <deployment_name>
{: pre}
kubectl rollout resume deployment <deployment_name>
{: pre}
-
Roll back a change.
-
View the roll-out history for the deployment and identify the revision number of your last deployment.
kubectl rollout history deployment/<deployment_name>
{: pre}
Tip: To see the details for a specific revision, include the revision number.
kubectl rollout history deployment/<deployment_name> --revision=<number>
{: pre}
-
Roll back to the previous version, or specify a revision. To roll back to the previous version, use the following command.
kubectl rollout undo deployment/<depoyment_name> --to-revision=<number>
{: pre}
-
{: #app_cicd}
With {{site.data.keyword.cloud_notm}} and other open source tools, you can set up continuous integration and delivery (CI/CD), version control, tool chains, and more to help automate app development and deployment. {: shortdesc}
To see a list of supported integrations and steps for setting up a continuous delivery pipeline, see Setting up continuous integration and delivery.
{: #copy_apps_cluster}
When you use a version control system such as Git, configuration management projects such as kustomize
, or continuous delivery tools such as Razee{: external} in your cluster, you can deploy your app configuration files quickly from cluster to cluster. Sometimes you have only a few deployments that you tested in a cluster and prefer to copy these deployments and redeploy in another cluster.
{: shortdesc}
Before you begin, you need two clusters and the Manager service access role for all namespaces in both clusters so that you can copy all the resources from one cluster and deploy them to another.
-
List all the configuration files in your cluster and verify that you want to copy these configurations.
kubectl get all
{: pre}
Example output
NAME READY STATUS RESTARTS AGE pod/java-web-6955bdbcdf-l756b 1/1 Running 0 59d NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE service/java-web NodePort 172.21.xxx.xxx <none> 8080:30889/TCP 59d NAME READY UP-TO-DATE AVAILABLE AGE deployment.apps/java-web 1/1 1 1 59d NAME DESIRED CURRENT READY AGE replicaset.apps/java-web-6955bdbcdf 1 1 1 59d
{: screen}
-
Copy the configuration files in your cluster to a local directory. The
--export
option removes cluster-specific information from the configuration files.kubectl get all -o yaml --export > myconfigs.yaml
{: pre}
-
Optional: If your cluster used multiple namespaces, create the same namespaces in the standard cluster and copy the image pull secret to each namespace.
-
Deploy the copied configuration files to your cluster. If a configuration file has specific information that can't be applied, you might need to update the configuration file and reapply.
kubectl apply -f myconfigs.yaml
{: pre}
Example output
pod/java-web-6955bdbcdf-l756b created service/java-web created deployment.apps/java-web created replicaset.apps/java-web-6955bdbcdf created
{: screen}
-
Verify that your configuration files are applied.
kubectl get all
{: pre}