che-operator/pkg/deploy/deployment.go

527 lines
18 KiB
Go

//
// Copyright (c) 2019-2023 Red Hat, Inc.
// This program and the accompanying materials are made
// available under the terms of the Eclipse Public License 2.0
// which is available at https://www.eclipse.org/legal/epl-2.0/
//
// SPDX-License-Identifier: EPL-2.0
//
// Contributors:
// Red Hat, Inc. - initial API and implementation
//
package deploy
import (
"context"
"fmt"
"sort"
"strings"
k8shelper "github.com/eclipse-che/che-operator/pkg/common/k8s-helper"
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
"k8s.io/apimachinery/pkg/labels"
"k8s.io/apimachinery/pkg/selection"
"k8s.io/utils/pointer"
"sigs.k8s.io/controller-runtime/pkg/client"
"github.com/devfile/devworkspace-operator/pkg/infrastructure"
chev2 "github.com/eclipse-che/che-operator/api/v2"
"github.com/eclipse-che/che-operator/pkg/common/chetypes"
"github.com/eclipse-che/che-operator/pkg/common/constants"
"github.com/eclipse-che/che-operator/pkg/common/test"
"github.com/eclipse-che/che-operator/pkg/common/utils"
"github.com/google/go-cmp/cmp"
"github.com/google/go-cmp/cmp/cmpopts"
"github.com/sirupsen/logrus"
appsv1 "k8s.io/api/apps/v1"
corev1 "k8s.io/api/core/v1"
"k8s.io/apimachinery/pkg/api/resource"
)
var DefaultDeploymentDiffOpts = cmp.Options{
cmpopts.IgnoreFields(appsv1.Deployment{}, "TypeMeta", "ObjectMeta", "Status"),
cmpopts.IgnoreFields(appsv1.DeploymentSpec{}, "Replicas", "RevisionHistoryLimit", "ProgressDeadlineSeconds"),
cmpopts.IgnoreFields(appsv1.DeploymentStrategy{}, "RollingUpdate"),
cmpopts.IgnoreFields(corev1.Container{}, "ReadinessProbe", "LivenessProbe", "TerminationMessagePath", "TerminationMessagePolicy", "SecurityContext"),
cmpopts.IgnoreFields(corev1.PodSpec{}, "DNSPolicy", "SchedulerName", "SecurityContext", "DeprecatedServiceAccount"),
cmpopts.IgnoreFields(corev1.ConfigMapVolumeSource{}, "DefaultMode"),
cmpopts.IgnoreFields(corev1.SecretVolumeSource{}, "DefaultMode"),
cmpopts.IgnoreFields(corev1.VolumeSource{}, "EmptyDir"),
cmp.Comparer(func(x, y resource.Quantity) bool {
return x.Cmp(y) == 0
}),
}
func SyncDeploymentSpecToCluster(
deployContext *chetypes.DeployContext,
deploymentSpec *appsv1.Deployment,
deploymentDiffOpts cmp.Options) (bool, error) {
if err := MountSecrets(deploymentSpec, deployContext); err != nil {
return false, err
}
if err := MountConfigMaps(deploymentSpec, deployContext); err != nil {
return false, err
}
if err := setDesiredReplicas(deploymentSpec, deployContext); err != nil {
return false, err
}
done, err := Sync(deployContext, deploymentSpec, deploymentDiffOpts)
if err != nil || !done {
// Failed to sync (update), let's delete and create instead
if err != nil && strings.Contains(err.Error(), "field is immutable") {
if _, err := DeleteNamespacedObject(deployContext, deploymentSpec.Name, &appsv1.Deployment{}); err != nil {
return false, err
}
// Deleted successfully, return original error
return false, err
}
return false, err
}
// always return true for tests
if test.IsTestMode() {
return true, nil
}
actual := &appsv1.Deployment{}
exists, err := GetNamespacedObject(deployContext, deploymentSpec.ObjectMeta.Name, actual)
if !exists || err != nil {
return false, err
}
provisioned := actual.Status.UnavailableReplicas == 0
if actual.Spec.Strategy.Type == appsv1.RollingUpdateDeploymentStrategyType && !provisioned {
logrus.Infof("Deployment %s is in the rolling update state.", deploymentSpec.Name)
}
return provisioned, nil
}
// OverrideDeployment with custom settings
func OverrideDeployment(
ctx *chetypes.DeployContext,
deployment *appsv1.Deployment,
overrideDeploymentSettings *chev2.Deployment) error {
for index, _ := range deployment.Spec.Template.Spec.Containers {
var overrideContainerSettings *chev2.Container
container := &deployment.Spec.Template.Spec.Containers[index]
if overrideDeploymentSettings != nil && len(overrideDeploymentSettings.Containers) > 0 {
if len(deployment.Spec.Template.Spec.Containers) == 1 {
overrideContainerSettings = &overrideDeploymentSettings.Containers[0]
} else {
for _, c := range overrideDeploymentSettings.Containers {
if c.Name == container.Name {
overrideContainerSettings = &c
break
}
}
}
}
if err := OverrideContainer(ctx.CheCluster.Namespace, container, overrideContainerSettings); err != nil {
return err
}
}
if !infrastructure.IsOpenShift() {
if overrideDeploymentSettings != nil {
if overrideDeploymentSettings.SecurityContext != nil {
if deployment.Spec.Template.Spec.SecurityContext == nil {
deployment.Spec.Template.Spec.SecurityContext = &corev1.PodSecurityContext{}
}
if overrideDeploymentSettings.SecurityContext.FsGroup != nil {
deployment.Spec.Template.Spec.SecurityContext.FSGroup = pointer.Int64Ptr(*overrideDeploymentSettings.SecurityContext.FsGroup)
}
if overrideDeploymentSettings.SecurityContext.RunAsUser != nil {
deployment.Spec.Template.Spec.SecurityContext.RunAsUser = pointer.Int64Ptr(*overrideDeploymentSettings.SecurityContext.RunAsUser)
}
}
}
}
return nil
}
func OverrideContainer(
namespace string,
container *corev1.Container,
overrideSettings *chev2.Container) error {
// Set empty CPU limits when possible:
// 1. If there is no LimitRange in the namespace
// 2. CPU limits is not overridden
// See details at https://github.com/eclipse/che/issues/22198
if overrideSettings == nil || overrideSettings.Resources == nil || overrideSettings.Resources.Limits == nil || overrideSettings.Resources.Limits.Cpu == nil {
// use NonCachedClient to avoid cache LimitRange objects
if limitRanges, err := k8shelper.New().GetClientset().CoreV1().LimitRanges(namespace).List(context.TODO(), metav1.ListOptions{}); err != nil {
return err
} else if len(limitRanges.Items) == 0 { // no LimitRange in the namespace
delete(container.Resources.Limits, corev1.ResourceCPU)
}
}
if overrideSettings == nil {
return nil
}
// Image
container.Image = utils.GetValue(overrideSettings.Image, container.Image)
container.ImagePullPolicy = overrideSettings.ImagePullPolicy
if container.ImagePullPolicy == "" {
container.ImagePullPolicy = corev1.PullPolicy(utils.GetPullPolicyFromDockerImage(container.Image))
}
// Env
for _, env := range overrideSettings.Env {
index := utils.IndexEnv(env.Name, container.Env)
if index == -1 {
container.Env = append(container.Env, env)
} else {
container.Env[index] = env
}
}
// Resources
if overrideSettings.Resources != nil {
if overrideSettings.Resources.Requests != nil {
if overrideSettings.Resources.Requests.Memory != nil {
if overrideSettings.Resources.Requests.Memory.IsZero() {
delete(container.Resources.Requests, corev1.ResourceMemory)
} else {
container.Resources.Requests[corev1.ResourceMemory] = *overrideSettings.Resources.Requests.Memory
}
}
if overrideSettings.Resources.Requests.Cpu != nil {
if overrideSettings.Resources.Requests.Cpu.IsZero() {
delete(container.Resources.Requests, corev1.ResourceCPU)
} else {
container.Resources.Requests[corev1.ResourceCPU] = *overrideSettings.Resources.Requests.Cpu
}
}
if len(container.Resources.Requests) == 0 {
container.Resources.Requests = nil
}
}
if overrideSettings.Resources.Limits != nil {
if overrideSettings.Resources.Limits.Memory != nil {
if overrideSettings.Resources.Limits.Memory.IsZero() {
delete(container.Resources.Limits, corev1.ResourceMemory)
} else {
container.Resources.Limits[corev1.ResourceMemory] = *overrideSettings.Resources.Limits.Memory
}
}
if overrideSettings.Resources.Limits.Cpu != nil {
if overrideSettings.Resources.Limits.Cpu.IsZero() {
delete(container.Resources.Limits, corev1.ResourceCPU)
} else {
container.Resources.Limits[corev1.ResourceCPU] = *overrideSettings.Resources.Limits.Cpu
}
}
if len(container.Resources.Limits) == 0 {
container.Resources.Limits = nil
}
}
}
return nil
}
// EnsurePodSecurityStandards sets SecurityContext accordingly
// to standards https://kubernetes.io/docs/concepts/security/pod-security-standards/#restricted
func EnsurePodSecurityStandards(deployment *appsv1.Deployment, userId int64, groupId int64) {
for i, _ := range deployment.Spec.Template.Spec.Containers {
if deployment.Spec.Template.Spec.Containers[i].SecurityContext == nil {
deployment.Spec.Template.Spec.Containers[i].SecurityContext = &corev1.SecurityContext{}
}
deployment.Spec.Template.Spec.Containers[i].SecurityContext.AllowPrivilegeEscalation = pointer.BoolPtr(false)
deployment.Spec.Template.Spec.Containers[i].SecurityContext.Capabilities = &corev1.Capabilities{Drop: []corev1.Capability{"ALL"}}
}
if !infrastructure.IsOpenShift() {
if deployment.Spec.Template.Spec.SecurityContext == nil {
deployment.Spec.Template.Spec.SecurityContext = &corev1.PodSecurityContext{}
}
deployment.Spec.Template.Spec.SecurityContext.RunAsUser = pointer.Int64Ptr(userId)
deployment.Spec.Template.Spec.SecurityContext.FSGroup = pointer.Int64Ptr(groupId)
}
}
// MountSecrets mounts secrets into a container as a file or as environment variable.
// Secrets are selected by the following labels:
// - app.kubernetes.io/part-of=che.eclipse.org
// - app.kubernetes.io/component=<DEPLOYMENT-NAME>-secret
func MountSecrets(specDeployment *appsv1.Deployment, deployContext *chetypes.DeployContext) error {
secrets := &corev1.SecretList{}
kubernetesPartOfLabelSelectorRequirement, _ := labels.NewRequirement(constants.KubernetesPartOfLabelKey, selection.Equals, []string{constants.CheEclipseOrg})
kubernetesComponentLabelSelectorRequirement, _ := labels.NewRequirement(constants.KubernetesComponentLabelKey, selection.Equals, []string{specDeployment.Name + "-secret"})
listOptions := &client.ListOptions{
LabelSelector: labels.NewSelector().Add(*kubernetesPartOfLabelSelectorRequirement).Add(*kubernetesComponentLabelSelectorRequirement),
}
if err := deployContext.ClusterAPI.Client.List(context.TODO(), secrets, listOptions); err != nil {
return err
}
// sort secrets by name first to have the same order every time
sort.Slice(secrets.Items, func(i, j int) bool {
return strings.Compare(secrets.Items[i].Name, secrets.Items[j].Name) < 0
})
container := &specDeployment.Spec.Template.Spec.Containers[0]
for _, secretObj := range secrets.Items {
switch secretObj.Annotations[constants.CheEclipseOrgMountAs] {
case "file":
volumeSource := corev1.VolumeSource{
Secret: &corev1.SecretVolumeSource{
SecretName: secretObj.Name,
},
}
volume := corev1.Volume{
Name: secretObj.Name,
VolumeSource: volumeSource,
}
specDeployment.Spec.Template.Spec.Volumes = append(specDeployment.Spec.Template.Spec.Volumes, volume)
volumeMount := corev1.VolumeMount{
Name: secretObj.Name,
MountPath: secretObj.Annotations[constants.CheEclipseOrgMountPath],
}
container.VolumeMounts = append(container.VolumeMounts, volumeMount)
case "subpath":
volumeSource := corev1.VolumeSource{
Secret: &corev1.SecretVolumeSource{
SecretName: secretObj.Name,
},
}
volume := corev1.Volume{
Name: secretObj.Name,
VolumeSource: volumeSource,
}
specDeployment.Spec.Template.Spec.Volumes = append(specDeployment.Spec.Template.Spec.Volumes, volume)
for fileName, _ := range secretObj.Data {
mountPath := secretObj.Annotations[constants.CheEclipseOrgMountPath]
if strings.HasSuffix(mountPath, "/") {
mountPath += fileName
} else {
mountPath += "/" + fileName
}
volumeMount := corev1.VolumeMount{
Name: secretObj.Name,
MountPath: mountPath,
SubPath: fileName,
}
container.VolumeMounts = append(container.VolumeMounts, volumeMount)
}
case "env":
secret := &corev1.Secret{}
exists, err := GetNamespacedObject(deployContext, secretObj.Name, secret)
if err != nil {
return err
} else if !exists {
return fmt.Errorf("Secret '%s' not found", secretObj.Name)
}
// grab all keys and sort first to have the same order every time
keys := make([]string, 0)
for k := range secret.Data {
keys = append(keys, k)
}
sort.Slice(keys, func(i, j int) bool {
return strings.Compare(keys[i], keys[j]) < 0
})
for _, key := range keys {
var envName string
// check if evn name defined explicitly
envNameAnnotation := constants.CheEclipseOrg + "/" + key + "_env-name"
envName, envNameExists := secretObj.Annotations[envNameAnnotation]
if !envNameExists {
// check if there is only one env name to mount
envName, envNameExists = secretObj.Annotations[constants.CheEclipseOrgEnvName]
if len(secret.Data) > 1 {
return fmt.Errorf("There are more than one environment variable to mount. Use annotation '%s' to specify a name", envNameAnnotation)
} else if !envNameExists {
return fmt.Errorf("Environment name to mount secret key not found. Use annotation '%s' to specify a name", constants.CheEclipseOrgEnvName)
}
}
env := corev1.EnvVar{
Name: envName,
ValueFrom: &corev1.EnvVarSource{
SecretKeyRef: &corev1.SecretKeySelector{
Key: key,
LocalObjectReference: corev1.LocalObjectReference{
Name: secretObj.Name,
},
},
},
}
container.Env = append(container.Env, env)
}
}
}
return nil
}
// MountConfigMaps mounts configmaps into a container as a file or as environment variable.
// Configmaps are selected by the following labels:
// - app.kubernetes.io/part-of=che.eclipse.org
// - app.kubernetes.io/component=<DEPLOYMENT-NAME>-configmap
func MountConfigMaps(specDeployment *appsv1.Deployment, deployContext *chetypes.DeployContext) error {
configmaps := &corev1.ConfigMapList{}
kubernetesPartOfLabelSelectorRequirement, _ := labels.NewRequirement(constants.KubernetesPartOfLabelKey, selection.Equals, []string{constants.CheEclipseOrg})
kubernetesComponentLabelSelectorRequirement, _ := labels.NewRequirement(constants.KubernetesComponentLabelKey, selection.Equals, []string{specDeployment.Name + "-configmap"})
listOptions := &client.ListOptions{
LabelSelector: labels.NewSelector().Add(*kubernetesPartOfLabelSelectorRequirement).Add(*kubernetesComponentLabelSelectorRequirement),
}
if err := deployContext.ClusterAPI.Client.List(context.TODO(), configmaps, listOptions); err != nil {
return err
}
// sort configmaps by name first to have the same order every time
sort.Slice(configmaps.Items, func(i, j int) bool {
return strings.Compare(configmaps.Items[i].Name, configmaps.Items[j].Name) < 0
})
container := &specDeployment.Spec.Template.Spec.Containers[0]
for _, configMapObj := range configmaps.Items {
switch configMapObj.Annotations[constants.CheEclipseOrgMountAs] {
case "file":
voluseSource := corev1.VolumeSource{
ConfigMap: &corev1.ConfigMapVolumeSource{
LocalObjectReference: corev1.LocalObjectReference{
Name: configMapObj.Name,
},
},
}
volume := corev1.Volume{
Name: configMapObj.Name,
VolumeSource: voluseSource,
}
volumeMount := corev1.VolumeMount{
Name: configMapObj.Name,
MountPath: configMapObj.Annotations[constants.CheEclipseOrgMountPath],
}
specDeployment.Spec.Template.Spec.Volumes = append(specDeployment.Spec.Template.Spec.Volumes, volume)
container.VolumeMounts = append(container.VolumeMounts, volumeMount)
case "subpath":
volumeSource := corev1.VolumeSource{
ConfigMap: &corev1.ConfigMapVolumeSource{
LocalObjectReference: corev1.LocalObjectReference{
Name: configMapObj.Name,
},
},
}
volume := corev1.Volume{
Name: configMapObj.Name,
VolumeSource: volumeSource,
}
specDeployment.Spec.Template.Spec.Volumes = append(specDeployment.Spec.Template.Spec.Volumes, volume)
for fileName, _ := range configMapObj.Data {
mountPath := configMapObj.Annotations[constants.CheEclipseOrgMountPath]
if strings.HasSuffix(mountPath, "/") {
mountPath += fileName
} else {
mountPath += "/" + fileName
}
volumeMount := corev1.VolumeMount{
Name: configMapObj.Name,
MountPath: mountPath,
SubPath: fileName,
}
container.VolumeMounts = append(container.VolumeMounts, volumeMount)
}
case "env":
configmap := &corev1.ConfigMap{}
exists, err := GetNamespacedObject(deployContext, configMapObj.Name, configmap)
if err != nil {
return err
} else if !exists {
return fmt.Errorf("ConfigMap '%s' not found", configMapObj.Name)
}
// grab all keys and sort first to have the same order every time
keys := make([]string, 0)
for k := range configmap.Data {
keys = append(keys, k)
}
sort.Slice(keys, func(i, j int) bool {
return strings.Compare(keys[i], keys[j]) < 0
})
for _, key := range keys {
var envName string
// check if evn name defined explicitly
envNameAnnotation := constants.CheEclipseOrg + "/" + key + "_env-name"
envName, envNameExists := configMapObj.Annotations[envNameAnnotation]
if !envNameExists {
// check if there is only one env name to mount
envName, envNameExists = configMapObj.Annotations[constants.CheEclipseOrgEnvName]
if len(configmap.Data) > 1 {
return fmt.Errorf("There are more than one environment variable to mount. Use annotation '%s' to specify a name", envNameAnnotation)
} else if !envNameExists {
return fmt.Errorf("Environment name to mount configmap key not found. Use annotation '%s' to specify a name", constants.CheEclipseOrgEnvName)
}
}
env := corev1.EnvVar{
Name: envName,
ValueFrom: &corev1.EnvVarSource{
ConfigMapKeyRef: &corev1.ConfigMapKeySelector{
Key: key,
LocalObjectReference: corev1.LocalObjectReference{
Name: configMapObj.Name,
},
},
},
}
container.Env = append(container.Env, env)
}
}
}
return nil
}
// setDesiredReplicas sets replicas count from the actual deployment.
func setDesiredReplicas(deployment *appsv1.Deployment, deployCtx *chetypes.DeployContext) error {
actual := &appsv1.Deployment{}
if exists, err := GetNamespacedObject(deployCtx, deployment.ObjectMeta.Name, actual); !exists {
return err
}
deployment.Spec.Replicas = actual.Spec.Replicas
return nil
}