From b72d97bd9716b142c6dab9c3bd63efcecf1bb425 Mon Sep 17 00:00:00 2001 From: emosbaugh Date: Wed, 27 Sep 2023 04:04:16 +0000 Subject: [PATCH] Create new OpenEBS version --- addons/openebs/3.9.0/Manifest | 2 + addons/openebs/3.9.0/install.sh | 449 ++++++++++++++++++ addons/openebs/3.9.0/spec/crds/crds.yaml | 383 +++++++++++++++ .../3.9.0/spec/crds/kustomization.yaml | 2 + addons/openebs/3.9.0/spec/openebs.tmpl.yaml | 178 +++++++ .../3.9.0/spec/storage/kustomization.yaml | 0 .../storage/tmpl-localpv-storage-class.yaml | 14 + .../storage/tmpl-patch-localpv-default.yaml | 6 + .../3.9.0/spec/tmpl-kustomization.yaml | 5 + addons/openebs/3.9.0/spec/tmpl-namespace.yaml | 4 + web/src/installers/versions.js | 1 + 11 files changed, 1044 insertions(+) create mode 100644 addons/openebs/3.9.0/Manifest create mode 100644 addons/openebs/3.9.0/install.sh create mode 100644 addons/openebs/3.9.0/spec/crds/crds.yaml create mode 100644 addons/openebs/3.9.0/spec/crds/kustomization.yaml create mode 100644 addons/openebs/3.9.0/spec/openebs.tmpl.yaml create mode 100644 addons/openebs/3.9.0/spec/storage/kustomization.yaml create mode 100644 addons/openebs/3.9.0/spec/storage/tmpl-localpv-storage-class.yaml create mode 100644 addons/openebs/3.9.0/spec/storage/tmpl-patch-localpv-default.yaml create mode 100644 addons/openebs/3.9.0/spec/tmpl-kustomization.yaml create mode 100644 addons/openebs/3.9.0/spec/tmpl-namespace.yaml diff --git a/addons/openebs/3.9.0/Manifest b/addons/openebs/3.9.0/Manifest new file mode 100644 index 0000000000..01b551a844 --- /dev/null +++ b/addons/openebs/3.9.0/Manifest @@ -0,0 +1,2 @@ +image linux-utils openebs/linux-utils:3.4.0 +image provisioner-localpv openebs/provisioner-localpv:3.4.0 diff --git a/addons/openebs/3.9.0/install.sh b/addons/openebs/3.9.0/install.sh new file mode 100644 index 0000000000..5c599e5881 --- /dev/null +++ b/addons/openebs/3.9.0/install.sh @@ -0,0 +1,449 @@ +# shellcheck disable=SC2148 + +function openebs_pre_init() { + if [ -z "$OPENEBS_NAMESPACE" ]; then + OPENEBS_NAMESPACE=openebs + fi + if [ -z "$OPENEBS_LOCALPV_STORAGE_CLASS" ]; then + OPENEBS_LOCALPV_STORAGE_CLASS=openebs-localpv + fi + if [ "$OPENEBS_CSTOR" = "1" ]; then + bail "cstor is not supported on OpenEBS $OPENEBS_APP_VERSION." + fi + + export OPENEBS_APP_VERSION="3.9.0" + export PREVIOUS_OPENEBS_VERSION="$(openebs_get_running_version)" + + openebs_bail_unsupported_upgrade + openebs_prompt_migrate_from_rook + openebs_prompt_migrate_from_longhorn +} + +function openebs() { + local src="$DIR/addons/openebs/$OPENEBS_VERSION" + local dst="$DIR/kustomize/openebs" + + secure_openebs + + openebs_apply_crds + + # migrate resources that are changing names + openebs_migrate_pre_helm_resources + + openebs_apply_operator + + # migrate resources that are changing names + openebs_migrate_post_helm_resources + + openebs_apply_storageclasses + + # if there is a validatingWebhookConfiguration, wait for the service to be ready + openebs_await_admissionserver + + # migrate from Rook/Ceph storage if applicable + openebs_maybe_migrate_from_rook + + # migrate from Longhorn storage if applicable + openebs_maybe_migrate_from_longhorn + + # remove NDM pods if applicable + openebs_cleanup_ndm +} + +# if rook-ceph is installed but is not specified in the kURL spec, migrate data from +# rook-ceph to OpenEBS local pv hostpath +function openebs_maybe_migrate_from_rook() { + if [ -z "$ROOK_VERSION" ]; then + if kubectl get ns | grep -q rook-ceph; then + # show validation errors from pvmigrate + # if there are errors, openebs_maybe_rook_migration_checks() will bail + openebs_maybe_rook_migration_checks + rook_ceph_to_sc_migration "$OPENEBS_LOCALPV_STORAGE_CLASS" "1" + # used to automatically delete rook-ceph if object store data was also migrated + add_rook_pvc_migration_status # used to automatically delete rook-ceph if object store data was also migrated + fi + fi +} + +function openebs_maybe_rook_migration_checks() { + logStep "Running Rook to OpenEBS migration checks ..." + + if ! rook_is_healthy_to_migrate_from; then + bail "Cannot upgrade from Rook to OpenEBS. Rook Ceph is unhealthy." + fi + + log "Awaiting 2 minutes to check OpenEBS Pod(s) are Running" + if ! spinner_until 120 check_for_running_pods "$OPENEBS_NAMESPACE"; then + logFail "OpenEBS has unhealthy Pod(s). Check the namespace $OPENEBS_NAMESPACE " + bail "Cannot upgrade from Rook to OpenEBS. OpenEBS is unhealthy." + fi + + # get the list of StorageClasses that use rook-ceph + local rook_scs + local rook_default_sc + rook_scs=$(kubectl get storageclass | grep rook | grep -v '(default)' | awk '{ print $1}') # any non-default rook StorageClasses + rook_default_sc=$(kubectl get storageclass | grep rook | grep '(default)' | awk '{ print $1}') # any default rook StorageClasses + + # Ensure openebs-localpv-provisioner deployment is ready + log "awaiting openebs-localpv-provisioner deployment" + spinner_until 120 deployment_fully_updated openebs openebs-localpv-provisioner + + local rook_scs_pvmigrate_dryrun_output + local rook_default_sc_pvmigrate_dryrun_output + for rook_sc in $rook_scs + do + # run validation checks for non default Rook storage classes + if ! rook_scs_pvmigrate_dryrun_output=$($BIN_PVMIGRATE --source-sc "$rook_sc" --dest-sc "$OPENEBS_LOCALPV_STORAGE_CLASS" --rsync-image "$KURL_UTIL_IMAGE" --preflight-validation-only) ; then + break + fi + done + + if [ -n "$rook_default_sc" ] ; then + # run validation checks for Rook default storage class + rook_default_sc_pvmigrate_dryrun_output=$($BIN_PVMIGRATE --source-sc "$rook_default_sc" --dest-sc "$OPENEBS_LOCALPV_STORAGE_CLASS" --rsync-image "$KURL_UTIL_IMAGE" --preflight-validation-only || true) + fi + + if [ -n "$rook_scs_pvmigrate_dryrun_output" ] || [ -n "$rook_default_sc_pvmigrate_dryrun_output" ] ; then + log "$rook_scs_pvmigrate_dryrun_output" + log "$rook_default_sc_pvmigrate_dryrun_output" + bail "Cannot upgrade from Rook to OpenEBS due to previous error." + fi + + logSuccess "Rook to OpenEBS migration checks completed successfully." +} + +function openebs_prompt_migrate_from_rook() { + local ceph_disk_usage_total + local rook_ceph_exec_deploy=rook-ceph-operator + + # skip on new install or when Rook is specified in the kURL spec + if [ -z "$CURRENT_KUBERNETES_VERSION" ] || [ -n "$ROOK_VERSION" ]; then + return 0 + fi + + # do not proceed if Rook is not installed + if ! kubectl get ns | grep -q rook-ceph; then + return 0 + fi + + if kubectl get deployment -n rook-ceph rook-ceph-tools &>/dev/null; then + rook_ceph_exec_deploy=rook-ceph-tools + fi + ceph_disk_usage_total=$(kubectl exec -n rook-ceph deployment/$rook_ceph_exec_deploy -- ceph df | grep TOTAL | awk '{ print $8$9 }') + + printf "${YELLOW}" + printf "\n" + printf " Detected Rook is running in the cluster. Data migration will be initiated to move data from rook-ceph to storage class %s.\n" "$OPENEBS_LOCALPV_STORAGE_CLASS" + printf "\n" + printf " As part of this, all pods mounting PVCs will be stopped, taking down the application.\n" + printf "\n" + printf " Copying the data currently stored within rook-ceph will require at least %s of free space across the cluster.\n" "$ceph_disk_usage_total" + printf " It is recommended to take a snapshot or otherwise back up your data before proceeding.\n${NC}" + printf "\n" + printf "Would you like to continue? " + + if ! confirmN; then + bail "Not migrating" + fi + + semverParse "$KUBERNETES_VERSION" + if [ "$minor" -gt 18 ] ; then + # if the current version of k8s is compatible with OpenEBS, install it and migrate the data before upgrading k8s + printf " Starting OpenEBS installation and migration from Rook to OpenEBS.\n" + + report_addon_start "openebs-preinstall" "$OPENEBS_VERSION" + addon_load "openebs" "$OPENEBS_VERSION" + addon_install "openebs" "$OPENEBS_VERSION" + report_addon_success "openebs-preinstall" "$OPENEBS_VERSION" + fi +} + +function openebs_apply_crds() { + local src="$DIR/addons/openebs/$OPENEBS_VERSION/spec/crds" + local dst="$DIR/kustomize/openebs/spec/crds" + + mkdir -p "$dst" + + cp "$src/kustomization.yaml" "$dst/" + cp "$src/crds.yaml" "$dst/" + + kubectl apply -k "$dst/" +} + +function openebs_apply_operator() { + local src="$DIR/addons/openebs/$OPENEBS_VERSION/spec" + local dst="$DIR/kustomize/openebs/spec" + + mkdir -p "$dst" + + render_yaml_file_2 "$src/tmpl-kustomization.yaml" > "$dst/kustomization.yaml" + render_yaml_file_2 "$src/tmpl-namespace.yaml" > "$dst/namespace.yaml" + cat "$src/openebs.tmpl.yaml" | sed "s/__OPENEBS_NAMESPACE__/$OPENEBS_NAMESPACE/" > "$dst/openebs.yaml" + + kubectl apply -k "$dst/" + + logStep "Waiting for OpenEBS CustomResourceDefinitions to be ready" + spinner_until 120 kubernetes_resource_exists default crd blockdevices.openebs.io + + openebs_cleanup_kubesystem + logSuccess "OpenEBS CustomResourceDefinitions are ready" +} + +function openebs_apply_storageclasses() { + # allow vendor to add custom storageclasses rather than the ones built into add-on + if [ "$OPENEBS_LOCALPV" != "1" ]; then + return + fi + + local src="$DIR/addons/openebs/$OPENEBS_VERSION/spec/storage" + local dst="$DIR/kustomize/openebs/spec/storage" + + mkdir -p "$dst" + + cp "$src/kustomization.yaml" "$dst/" + + if [ "$OPENEBS_LOCALPV" = "1" ]; then + report_addon_start "openebs-localpv" "$OPENEBS_APP_VERSION" + + render_yaml_file_2 "$src/tmpl-localpv-storage-class.yaml" > "$dst/localpv-storage-class.yaml" + insert_resources "$dst/kustomization.yaml" localpv-storage-class.yaml + + if openebs_should_be_default_storageclass "$OPENEBS_LOCALPV_STORAGE_CLASS" ; then + echo "OpenEBS LocalPV will be installed as the default storage class." + render_yaml_file_2 "$src/tmpl-patch-localpv-default.yaml" > "$dst/patch-localpv-default.yaml" + insert_patches_strategic_merge "$dst/kustomization.yaml" patch-localpv-default.yaml + else + logWarn "Existing default storage class that is not OpenEBS LocalPV detected." + logWarn "OpenEBS LocalPV will be installed as the non-default storage class." + fi + + report_addon_success "openebs-localpv" "$OPENEBS_APP_VERSION" + fi + + kubectl apply -k "$dst/" +} + +function openebs_await_admissionserver() { + sleep 1 + if kubectl get validatingWebhookConfiguration openebs-validation-webhook-cfg &>/dev/null ; then + logStep "Waiting for OpenEBS admission controller service to be ready" + spinner_until 120 kubernetes_service_healthy "$OPENEBS_NAMESPACE" admission-server-svc + logSuccess "OpenEBS admission controller service is ready" + fi +} + +function openebs_join() { + secure_openebs +} + +function openebs_get_running_version() { + if kubectl get ns "$OPENEBS_NAMESPACE" >/dev/null 2>&1 ; then + kubectl -n "$OPENEBS_NAMESPACE" get deploy openebs-provisioner -o jsonpath='{.metadata.labels.openebs\.io/version}' 2>/dev/null + fi +} + +# upgrading non-cstor openebs works from 1.x +function openebs_bail_unsupported_upgrade() { + if [ -z "$PREVIOUS_OPENEBS_VERSION" ]; then + return 0 + fi +} + +function secure_openebs() { + mkdir -p /var/openebs + chmod 700 /var/openebs +} + +function openebs_cleanup_kubesystem() { + # cleanup old kube-system statefulsets + # https://github.com/openebs/upgrade/blob/v2.12.2/docs/upgrade.md#prerequisites-1 + kubectl -n kube-system delete sts openebs-cstor-csi-controller 2>/dev/null || true + kubectl -n kube-system delete ds openebs-cstor-csi-node 2>/dev/null || true + kubectl -n kube-system delete sa openebs-cstor-csi-controller-sa,openebs-cstor-csi-node-sa 2>/dev/null || true +} + +function openebs_migrate_pre_helm_resources() { + # name changed from maya-apiserver-service > openebs-apiservice + kubectl -n "$OPENEBS_NAMESPACE" delete service maya-apiserver-service 2>/dev/null || true + # name changed from cvc-operator-service > openebs-cstor-cvc-operator-svc + kubectl -n "$OPENEBS_NAMESPACE" delete service cvc-operator-service 2>/dev/null || true + # name changed from maya-apiserver >openebs-apiserver + kubectl -n "$OPENEBS_NAMESPACE" delete deployment maya-apiserver 2>/dev/null || true + # name changed from cspc-operator > openebs-cstor-cspc-operator + kubectl -n "$OPENEBS_NAMESPACE" delete deployment cspc-operator 2>/dev/null || true + # name changed from cvc-operator > openebs-cstor-cvc-operator + kubectl -n "$OPENEBS_NAMESPACE" delete deployment cvc-operator 2>/dev/null || true + + # the selectors for these resources changed + kubectl -n "$OPENEBS_NAMESPACE" delete deployment openebs-localpv-provisioner 2>/dev/null || true + kubectl -n "$OPENEBS_NAMESPACE" delete deployment openebs-ndm-operator 2>/dev/null || true + kubectl -n "$OPENEBS_NAMESPACE" delete daemonset openebs-ndm 2>/dev/null || true + kubectl -n "$OPENEBS_NAMESPACE" delete deployment openebs-provisioner 2>/dev/null || true + kubectl -n "$OPENEBS_NAMESPACE" delete deployment openebs-snapshot-operator 2>/dev/null || true + + # cleanup admission webhook + kubectl delete validatingWebhookConfiguration openebs-validation-webhook-cfg 2>/dev/null || true + kubectl -n "$OPENEBS_NAMESPACE" delete deployment openebs-admission-server 2>/dev/null || true + kubectl -n "$OPENEBS_NAMESPACE" delete service admission-server-svc 2>/dev/null || true +} + +function openebs_migrate_post_helm_resources() { + # name changed from openebs-maya-operator > openebs + kubectl delete serviceaccount openebs-maya-operator 2>/dev/null || true + # name changed from openebs-maya-operator > openebs + kubectl delete clusterrole openebs-maya-operator 2>/dev/null || true + # name changed from openebs-maya-operator > openebs + kubectl delete clusterrolebinding openebs-maya-operator 2>/dev/null || true +} + +function openebs_should_be_default_storageclass() { + local storage_class_name="$1" + if openebs_is_default_storageclass "$storage_class_name" ; then + # if "$storage_class_name" is already the default + return 0 + elif openebs_has_default_storageclass ; then + # if there is already a default storage class that is not "$storage_class_name" + return 1 + elif [ -n "$ROOK_MINIMUM_NODE_COUNT" ] && [ "$ROOK_MINIMUM_NODE_COUNT" -gt "1" ]; then + # if dynamic storage is enabled, the default storageclass will be managed by ekco + return 1 + elif [ "$storage_class_name" = "default" ]; then + # if "$storage_class_name" named "default", it should be the default + return 0 + elif [ -n "$LONGHORN_VERSION" ]; then + # To maintain backwards compatibility with previous versions of kURL, only make OpenEBS the default + # if Longhorn is not installed or the storageclass is explicitly named "default" + return 1 + else + # if there is no other storageclass, make "$storage_class_name" the default + return 0 + fi +} + +function openebs_is_default_storageclass() { + local storage_class_name="$1" + if [ "$(kubectl get sc "$storage_class_name" -o jsonpath='{.metadata.annotations.storageclass\.kubernetes\.io/is-default-class}' 2>/dev/null)" = "true" ]; then + return 0 + fi + return 1 +} + +function openebs_has_default_storageclass() { + if kubectl get sc -o jsonpath='{.items[*].metadata.annotations.storageclass\.kubernetes\.io/is-default-class}' | grep -q "true" ; then + return 0 + fi + return 1 +} + +# if longhorn is installed but is not specified in the kURL spec, migrate data to OpenEBS local pv hostpath. +function openebs_maybe_migrate_from_longhorn() { + if [ -z "$LONGHORN_VERSION" ]; then + if kubectl get ns | grep -q longhorn-system; then + # show validation errors from pvmigrate if there are errors, openebs_maybe_longhorn_migration_checks() will bail + openebs_maybe_longhorn_migration_checks + + longhorn_to_sc_migration "$OPENEBS_LOCALPV_STORAGE_CLASS" "1" + DID_MIGRATE_LONGHORN_PVCS=1 # used to automatically delete longhorn if object store data was also migrated + fi + fi +} + +function openebs_maybe_longhorn_migration_checks() { + logStep "Running Longhorn to OpenEBS migration checks" + + log "Awaiting 2 minutes to check OpenEBS Pod(s) are Running" + if ! spinner_until 120 check_for_running_pods "$OPENEBS_NAMESPACE"; then + logFail "OpenEBS has unhealthy Pod(s). Check the namespace $OPENEBS_NAMESPACE " + bail "Cannot upgrade from Rook to OpenEBS. OpenEBS is unhealthy." + fi + + log "Awaiting 2 minutes to check Longhorn Pod(s) are Running" + if ! spinner_until 120 check_for_running_pods longhorn-system; then + logFail "Longhorn has unhealthy Pod(s). Check the namespace longhorn-system" + bail "Cannot upgrade from Longhorn to OpenEBS. Longhorn is unhealthy." + fi + + # get the list of StorageClasses that use longhorn + local longhorn_scs + local longhorn_default_sc + longhorn_scs=$(kubectl get storageclass | grep longhorn | grep -v '(default)' | awk '{ print $1}') # any non-default longhorn StorageClasses + longhorn_default_sc=$(kubectl get storageclass | grep longhorn | grep '(default)' | awk '{ print $1}') # any default longhorn StorageClasses + + # Ensure openebs-localpv-provisioner deployment is ready + log "Awaiting openebs-localpv-provisioner deployment" + spinner_until 120 deployment_fully_updated openebs openebs-localpv-provisioner + + local longhorn_scs_pvmigrate_dryrun_output + local longhorn_default_sc_pvmigrate_dryrun_output + for longhorn_sc in $longhorn_scs + do + # run validation checks for non default Longhorn storage classes + if longhorn_scs_pvmigrate_dryrun_output=$($BIN_PVMIGRATE --source-sc "$longhorn_sc" --dest-sc "$OPENEBS_LOCALPV_STORAGE_CLASS" --rsync-image "$KURL_UTIL_IMAGE" --preflight-validation-only 2>&1) ; then + longhorn_scs_pvmigrate_dryrun_output="" + else + break + fi + done + + if [ -n "$longhorn_default_sc" ] ; then + # run validation checks for Longhorn default storage class + if longhorn_default_sc_pvmigrate_dryrun_output=$($BIN_PVMIGRATE --source-sc "$longhorn_default_sc" --dest-sc "$OPENEBS_LOCALPV_STORAGE_CLASS" --rsync-image "$KURL_UTIL_IMAGE" --preflight-validation-only 2>&1) ; then + longhorn_default_sc_pvmigrate_dryrun_output="" + fi + fi + + if [ -n "$longhorn_scs_pvmigrate_dryrun_output" ] || [ -n "$longhorn_default_sc_pvmigrate_dryrun_output" ] ; then + log "$longhorn_scs_pvmigrate_dryrun_output" + log "$longhorn_default_sc_pvmigrate_dryrun_output" + longhorn_restore_migration_replicas + bail "Cannot upgrade from Longhorn to OpenEBS due to previous error." + fi + + logSuccess "Longhorn to OpenEBS migration checks completed." +} + +# shows a prompt asking users for confirmation before starting to migrate data from Longhorn. +function openebs_prompt_migrate_from_longhorn() { + # skip on new install or when Longhorn is specified in the kURL spec + if [ -z "$CURRENT_KUBERNETES_VERSION" ] || [ -n "$LONGHORN_VERSION" ]; then + return 0 + fi + + # do not proceed if Longhorn is not installed + if ! kubectl get ns | grep -q longhorn-system; then + return 0 + fi + + logWarn " Detected Longhorn is running in the cluster. Data migration will be initiated to move data from Longhorn to storage class $OPENEBS_LOCALPV_STORAGE_CLASS." + logWarn " As part of this, all pods mounting PVCs will be stopped, taking down the application." + logWarn " It is recommended to take a snapshot or otherwise back up your data before proceeding." + + log "Would you like to continue? " + if ! confirmN; then + bail "Not migrating" + fi + + semverParse "$KUBERNETES_VERSION" + if [ "$minor" -gt 18 ] ; then + # if the current version of k8s is compatible with OpenEBS, install it and migrate the data before upgrading k8s + printf " Starting OpenEBS installation and migration from Longhorn to OpenEBS.\n" + + longhorn_prepare_for_migration + report_addon_start "openebs-preinstall" "$OPENEBS_VERSION" + addon_load "openebs" "$OPENEBS_VERSION" + addon_install "openebs" "$OPENEBS_VERSION" + report_addon_success "openebs-preinstall" "$OPENEBS_VERSION" + maybe_cleanup_longhorn + return + fi + + if ! longhorn_prepare_for_migration; then + bail "Not migrating" + fi +} + +function openebs_cleanup_ndm() { + kubectl delete --ignore-not-found configmap -n "$OPENEBS_NAMESPACE" openebs-ndm-config + kubectl delete --ignore-not-found daemonset -n "$OPENEBS_NAMESPACE" openebs-ndm + kubectl delete --ignore-not-found deployment -n "$OPENEBS_NAMESPACE" openebs-ndm-operator +} diff --git a/addons/openebs/3.9.0/spec/crds/crds.yaml b/addons/openebs/3.9.0/spec/crds/crds.yaml new file mode 100644 index 0000000000..d06ef2472b --- /dev/null +++ b/addons/openebs/3.9.0/spec/crds/crds.yaml @@ -0,0 +1,383 @@ +apiVersion: apiextensions.k8s.io/v1 +kind: CustomResourceDefinition +metadata: + annotations: + controller-gen.kubebuilder.io/version: v0.5.0 + creationTimestamp: null + name: blockdevices.openebs.io +spec: + group: openebs.io + names: + kind: BlockDevice + listKind: BlockDeviceList + plural: blockdevices + shortNames: + - bd + singular: blockdevice + scope: Namespaced + versions: + - additionalPrinterColumns: + - jsonPath: .spec.nodeAttributes.nodeName + name: NodeName + type: string + - jsonPath: .spec.path + name: Path + priority: 1 + type: string + - jsonPath: .spec.filesystem.fsType + name: FSType + priority: 1 + type: string + - jsonPath: .spec.capacity.storage + name: Size + type: string + - jsonPath: .status.claimState + name: ClaimState + type: string + - jsonPath: .status.state + name: Status + type: string + - jsonPath: .metadata.creationTimestamp + name: Age + type: date + name: v1alpha1 + schema: + openAPIV3Schema: + description: BlockDevice is the Schema for the blockdevices API + properties: + apiVersion: + description: 'APIVersion defines the versioned schema of this representation of an object. Servers should convert recognized schemas to the latest internal value, and may reject unrecognized values. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources' + type: string + kind: + description: 'Kind is a string value representing the REST resource this object represents. Servers may infer this from the endpoint the client submits requests to. Cannot be updated. In CamelCase. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds' + type: string + metadata: + type: object + spec: + description: DeviceSpec defines the properties and runtime status of a BlockDevice + properties: + aggregateDevice: + description: AggregateDevice was intended to store the hierarchical information in cases of LVM. However this is currently not implemented and may need to be re-looked into for better design. To be deprecated + type: string + capacity: + description: Capacity + properties: + logicalSectorSize: + description: LogicalSectorSize is blockdevice logical-sector size in bytes + format: int32 + type: integer + physicalSectorSize: + description: PhysicalSectorSize is blockdevice physical-Sector size in bytes + format: int32 + type: integer + storage: + description: Storage is the blockdevice capacity in bytes + format: int64 + type: integer + required: + - storage + type: object + claimRef: + description: ClaimRef is the reference to the BDC which has claimed this BD + properties: + apiVersion: + description: API version of the referent. + type: string + fieldPath: + description: 'If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object. TODO: this design is not final and this field is subject to change in the future.' + type: string + kind: + description: 'Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds' + type: string + name: + description: 'Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names' + type: string + namespace: + description: 'Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/' + type: string + resourceVersion: + description: 'Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency' + type: string + uid: + description: 'UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids' + type: string + type: object + details: + description: Details contain static attributes of BD like model,serial, and so forth + properties: + compliance: + description: Compliance is standards/specifications version implemented by device firmware such as SPC-1, SPC-2, etc + type: string + deviceType: + description: DeviceType represents the type of device like sparse, disk, partition, lvm, crypt + enum: + - disk + - partition + - sparse + - loop + - lvm + - crypt + - dm + - mpath + type: string + driveType: + description: DriveType is the type of backing drive, HDD/SSD + enum: + - HDD + - SSD + - Unknown + - "" + type: string + firmwareRevision: + description: FirmwareRevision is the disk firmware revision + type: string + hardwareSectorSize: + description: HardwareSectorSize is the hardware sector size in bytes + format: int32 + type: integer + logicalBlockSize: + description: LogicalBlockSize is the logical block size in bytes reported by /sys/class/block/sda/queue/logical_block_size + format: int32 + type: integer + model: + description: Model is model of disk + type: string + physicalBlockSize: + description: PhysicalBlockSize is the physical block size in bytes reported by /sys/class/block/sda/queue/physical_block_size + format: int32 + type: integer + serial: + description: Serial is serial number of disk + type: string + vendor: + description: Vendor is vendor of disk + type: string + type: object + devlinks: + description: DevLinks contains soft links of a block device like /dev/by-id/... /dev/by-uuid/... + items: + description: DeviceDevLink holds the mapping between type and links like by-id type or by-path type link + properties: + kind: + description: Kind is the type of link like by-id or by-path. + enum: + - by-id + - by-path + type: string + links: + description: Links are the soft links + items: + type: string + type: array + type: object + type: array + filesystem: + description: FileSystem contains mountpoint and filesystem type + properties: + fsType: + description: Type represents the FileSystem type of the block device + type: string + mountPoint: + description: MountPoint represents the mountpoint of the block device. + type: string + type: object + nodeAttributes: + description: NodeAttributes has the details of the node on which BD is attached + properties: + nodeName: + description: NodeName is the name of the Kubernetes node resource on which the device is attached + type: string + type: object + parentDevice: + description: "ParentDevice was intended to store the UUID of the parent Block Device as is the case for partitioned block devices. \n For example: /dev/sda is the parent for /dev/sda1 To be deprecated" + type: string + partitioned: + description: Partitioned represents if BlockDevice has partitions or not (Yes/No) Currently always default to No. To be deprecated + enum: + - "Yes" + - "No" + type: string + path: + description: Path contain devpath (e.g. /dev/sdb) + type: string + required: + - capacity + - devlinks + - nodeAttributes + - path + type: object + status: + description: DeviceStatus defines the observed state of BlockDevice + properties: + claimState: + description: ClaimState represents the claim state of the block device + enum: + - Claimed + - Unclaimed + - Released + type: string + state: + description: State is the current state of the blockdevice (Active/Inactive/Unknown) + enum: + - Active + - Inactive + - Unknown + type: string + required: + - claimState + - state + type: object + type: object + served: true + storage: true + subresources: {} +status: + acceptedNames: + kind: "" + plural: "" + conditions: [] + storedVersions: [] + +--- +apiVersion: apiextensions.k8s.io/v1 +kind: CustomResourceDefinition +metadata: + annotations: + controller-gen.kubebuilder.io/version: v0.5.0 + creationTimestamp: null + name: blockdeviceclaims.openebs.io +spec: + group: openebs.io + names: + kind: BlockDeviceClaim + listKind: BlockDeviceClaimList + plural: blockdeviceclaims + shortNames: + - bdc + singular: blockdeviceclaim + scope: Namespaced + versions: + - additionalPrinterColumns: + - jsonPath: .spec.blockDeviceName + name: BlockDeviceName + type: string + - jsonPath: .status.phase + name: Phase + type: string + - jsonPath: .metadata.creationTimestamp + name: Age + type: date + name: v1alpha1 + schema: + openAPIV3Schema: + description: BlockDeviceClaim is the Schema for the blockdeviceclaims API + properties: + apiVersion: + description: 'APIVersion defines the versioned schema of this representation of an object. Servers should convert recognized schemas to the latest internal value, and may reject unrecognized values. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources' + type: string + kind: + description: 'Kind is a string value representing the REST resource this object represents. Servers may infer this from the endpoint the client submits requests to. Cannot be updated. In CamelCase. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds' + type: string + metadata: + type: object + spec: + description: DeviceClaimSpec defines the request details for a BlockDevice + properties: + blockDeviceName: + description: BlockDeviceName is the reference to the block-device backing this claim + type: string + blockDeviceNodeAttributes: + description: BlockDeviceNodeAttributes is the attributes on the node from which a BD should be selected for this claim. It can include nodename, failure domain etc. + properties: + hostName: + description: HostName represents the hostname of the Kubernetes node resource where the BD should be present + type: string + nodeName: + description: NodeName represents the name of the Kubernetes node resource where the BD should be present + type: string + type: object + deviceClaimDetails: + description: Details of the device to be claimed + properties: + allowPartition: + description: AllowPartition represents whether to claim a full block device or a device that is a partition + type: boolean + blockVolumeMode: + description: 'BlockVolumeMode represents whether to claim a device in Block mode or Filesystem mode. These are use cases of BlockVolumeMode: 1) Not specified: VolumeMode check will not be effective 2) VolumeModeBlock: BD should not have any filesystem or mountpoint 3) VolumeModeFileSystem: BD should have a filesystem and mountpoint. If DeviceFormat is specified then the format should match with the FSType in BD' + type: string + formatType: + description: Format of the device required, eg:ext4, xfs + type: string + type: object + deviceType: + description: DeviceType represents the type of drive like SSD, HDD etc., + nullable: true + type: string + hostName: + description: Node name from where blockdevice has to be claimed. To be deprecated. Use NodeAttributes.HostName instead + type: string + resources: + description: Resources will help with placing claims on Capacity, IOPS + properties: + requests: + additionalProperties: + anyOf: + - type: integer + - type: string + pattern: ^(\+|-)?(([0-9]+(\.[0-9]*)?)|(\.[0-9]+))(([KMGTPE]i)|[numkMGTPE]|([eE](\+|-)?(([0-9]+(\.[0-9]*)?)|(\.[0-9]+))))?$ + x-kubernetes-int-or-string: true + description: 'Requests describes the minimum resources required. eg: if storage resource of 10G is requested minimum capacity of 10G should be available TODO for validating' + type: object + required: + - requests + type: object + selector: + description: Selector is used to find block devices to be considered for claiming + properties: + matchExpressions: + description: matchExpressions is a list of label selector requirements. The requirements are ANDed. + items: + description: A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values. + properties: + key: + description: key is the label key that the selector applies to. + type: string + operator: + description: operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist. + type: string + values: + description: values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch. + items: + type: string + type: array + required: + - key + - operator + type: object + type: array + matchLabels: + additionalProperties: + type: string + description: matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. + type: object + type: object + type: object + status: + description: DeviceClaimStatus defines the observed state of BlockDeviceClaim + properties: + phase: + description: Phase represents the current phase of the claim + type: string + required: + - phase + type: object + type: object + served: true + storage: true + subresources: {} +status: + acceptedNames: + kind: "" + plural: "" + conditions: [] + storedVersions: [] diff --git a/addons/openebs/3.9.0/spec/crds/kustomization.yaml b/addons/openebs/3.9.0/spec/crds/kustomization.yaml new file mode 100644 index 0000000000..8be8c0b2aa --- /dev/null +++ b/addons/openebs/3.9.0/spec/crds/kustomization.yaml @@ -0,0 +1,2 @@ +resources: +- crds.yaml diff --git a/addons/openebs/3.9.0/spec/openebs.tmpl.yaml b/addons/openebs/3.9.0/spec/openebs.tmpl.yaml new file mode 100644 index 0000000000..ebbb9ec366 --- /dev/null +++ b/addons/openebs/3.9.0/spec/openebs.tmpl.yaml @@ -0,0 +1,178 @@ +# Source: openebs/templates/serviceaccount.yaml +apiVersion: v1 +kind: ServiceAccount +metadata: + name: openebs + labels: + app: openebs + chart: openebs-3.9.0 + release: openebs + heritage: Helm +--- +# Source: openebs/templates/clusterrole.yaml +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRole +metadata: + name: openebs + labels: + app: openebs + chart: openebs-3.9.0 + release: openebs + heritage: Helm +rules: +- apiGroups: ["*"] + resources: ["nodes", "nodes/proxy"] + verbs: ["*"] +- apiGroups: ["*"] + resources: ["namespaces", "services", "pods", "pods/exec", "deployments", "deployments/finalizers", "replicationcontrollers", "replicasets", "events", "endpoints", "configmaps", "secrets", "jobs", "cronjobs" ] + verbs: ["*"] +- apiGroups: ["*"] + resources: ["statefulsets", "daemonsets"] + verbs: ["*"] +- apiGroups: ["*"] + resources: ["resourcequotas", "limitranges"] + verbs: ["list", "watch"] +- apiGroups: ["*"] + resources: ["ingresses", "horizontalpodautoscalers", "verticalpodautoscalers", "poddisruptionbudgets", "certificatesigningrequests"] + verbs: ["list", "watch"] +- apiGroups: ["*"] + resources: ["storageclasses", "persistentvolumeclaims", "persistentvolumes"] + verbs: ["*"] +- apiGroups: ["volumesnapshot.external-storage.k8s.io"] + resources: ["volumesnapshots", "volumesnapshotdatas"] + verbs: ["get", "list", "watch", "create", "update", "patch", "delete"] +- apiGroups: ["apiextensions.k8s.io"] + resources: ["customresourcedefinitions"] + verbs: [ "get", "list", "create", "update", "delete", "patch"] +- apiGroups: ["openebs.io"] + resources: [ "*"] + verbs: ["*" ] +- apiGroups: ["cstor.openebs.io"] + resources: [ "*"] + verbs: ["*" ] +- apiGroups: ["coordination.k8s.io"] + resources: ["leases"] + verbs: ["get", "watch", "list", "delete", "update", "create"] +- apiGroups: ["admissionregistration.k8s.io"] + resources: ["validatingwebhookconfigurations", "mutatingwebhookconfigurations"] + verbs: ["get", "create", "list", "delete", "update", "patch"] +- nonResourceURLs: ["/metrics"] + verbs: ["get"] +--- +# Source: openebs/templates/clusterrolebinding.yaml +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRoleBinding +metadata: + name: openebs + labels: + app: openebs + chart: openebs-3.9.0 + release: openebs + heritage: Helm +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: ClusterRole + name: openebs +subjects: +- kind: ServiceAccount + name: openebs + namespace: __OPENEBS_NAMESPACE__ +--- +# Source: openebs/templates/localprovisioner/deployment-local-provisioner.yaml +apiVersion: apps/v1 +kind: Deployment +metadata: + name: openebs-localpv-provisioner + labels: + app: openebs + chart: openebs-3.9.0 + release: openebs + heritage: Helm + component: localpv-provisioner + openebs.io/component-name: openebs-localpv-provisioner + openebs.io/version: 3.9.0 +spec: + replicas: 1 + strategy: + type: "Recreate" + rollingUpdate: null + selector: + matchLabels: + app: openebs + release: openebs + template: + metadata: + labels: + app: openebs + release: openebs + component: localpv-provisioner + name: openebs-localpv-provisioner + openebs.io/component-name: openebs-localpv-provisioner + openebs.io/version: 3.9.0 + spec: + serviceAccountName: openebs + containers: + - name: openebs-localpv-provisioner + image: "openebs/provisioner-localpv:3.4.0" + imagePullPolicy: IfNotPresent + args: + - "--bd-time-out=$(BDC_BD_BIND_RETRIES)" + env: + # OPENEBS_IO_K8S_MASTER enables openebs provisioner to connect to K8s + # based on this address. This is ignored if empty. + # This is supported for openebs provisioner version 0.5.2 onwards + #- name: OPENEBS_IO_K8S_MASTER + # value: "http://10.128.0.12:8080" + # OPENEBS_IO_KUBE_CONFIG enables openebs provisioner to connect to K8s + # based on this config. This is ignored if empty. + # This is supported for openebs provisioner version 0.5.2 onwards + #- name: OPENEBS_IO_KUBE_CONFIG + # value: "/home/ubuntu/.kube/config" + # This sets the number of times the provisioner should try + # with a polling interval of 5 seconds, to get the Blockdevice + # Name from a BlockDeviceClaim, before the BlockDeviceClaim + # is deleted. E.g. 12 * 5 seconds = 60 seconds timeout + - name: BDC_BD_BIND_RETRIES + value: "12" + # OPENEBS_NAMESPACE is the namespace that this provisioner will + # lookup to find maya api service + - name: OPENEBS_NAMESPACE + value: "__OPENEBS_NAMESPACE__" + - name: NODE_NAME + valueFrom: + fieldRef: + fieldPath: spec.nodeName + # OPENEBS_SERVICE_ACCOUNT provides the service account of this pod as + # environment variable + - name: OPENEBS_SERVICE_ACCOUNT + valueFrom: + fieldRef: + fieldPath: spec.serviceAccountName + # OPENEBS_IO_BASE_PATH is the environment variable that provides the + # default base path on the node where host-path PVs will be provisioned. + - name: OPENEBS_IO_ENABLE_ANALYTICS + value: "true" + - name: OPENEBS_IO_BASE_PATH + value: "/var/openebs/local" + - name: OPENEBS_IO_HELPER_IMAGE + value: "openebs/linux-utils:3.4.0" + - name: OPENEBS_IO_INSTALLER_TYPE + value: "charts-helm" + # LEADER_ELECTION_ENABLED is used to enable/disable leader election. By default + # leader election is enabled. + - name: LEADER_ELECTION_ENABLED + value: "true" + # Process name used for matching is limited to the 15 characters + # present in the pgrep output. + # So fullname can't be used here with pgrep (>15 chars).A regular expression + # that matches the entire command name has to specified. + # Anchor `^` : matches any string that starts with `provisioner-loc` + # `.*`: matches any string that has `provisioner-loc` followed by zero or more char + livenessProbe: + exec: + command: + - sh + - -c + - test `pgrep -c "^provisioner-loc.*"` = 1 + initialDelaySeconds: 30 + periodSeconds: 60 \ No newline at end of file diff --git a/addons/openebs/3.9.0/spec/storage/kustomization.yaml b/addons/openebs/3.9.0/spec/storage/kustomization.yaml new file mode 100644 index 0000000000..e69de29bb2 diff --git a/addons/openebs/3.9.0/spec/storage/tmpl-localpv-storage-class.yaml b/addons/openebs/3.9.0/spec/storage/tmpl-localpv-storage-class.yaml new file mode 100644 index 0000000000..e49d0e0a6e --- /dev/null +++ b/addons/openebs/3.9.0/spec/storage/tmpl-localpv-storage-class.yaml @@ -0,0 +1,14 @@ +apiVersion: storage.k8s.io/v1 +kind: StorageClass +metadata: + name: ${OPENEBS_LOCALPV_STORAGE_CLASS} + annotations: + openebs.io/cas-type: local + cas.openebs.io/config: | + - name: BasePath + value: "/var/openebs/local" + - name: StorageType + value: "hostpath" +provisioner: openebs.io/local +reclaimPolicy: Delete +volumeBindingMode: WaitForFirstConsumer diff --git a/addons/openebs/3.9.0/spec/storage/tmpl-patch-localpv-default.yaml b/addons/openebs/3.9.0/spec/storage/tmpl-patch-localpv-default.yaml new file mode 100644 index 0000000000..9f9958e089 --- /dev/null +++ b/addons/openebs/3.9.0/spec/storage/tmpl-patch-localpv-default.yaml @@ -0,0 +1,6 @@ +apiVersion: storage.k8s.io/v1 +kind: StorageClass +metadata: + name: ${OPENEBS_LOCALPV_STORAGE_CLASS} + annotations: + storageclass.kubernetes.io/is-default-class: "true" diff --git a/addons/openebs/3.9.0/spec/tmpl-kustomization.yaml b/addons/openebs/3.9.0/spec/tmpl-kustomization.yaml new file mode 100644 index 0000000000..018a677eb9 --- /dev/null +++ b/addons/openebs/3.9.0/spec/tmpl-kustomization.yaml @@ -0,0 +1,5 @@ +namespace: ${OPENEBS_NAMESPACE} + +resources: +- namespace.yaml +- openebs.yaml diff --git a/addons/openebs/3.9.0/spec/tmpl-namespace.yaml b/addons/openebs/3.9.0/spec/tmpl-namespace.yaml new file mode 100644 index 0000000000..f9182714b9 --- /dev/null +++ b/addons/openebs/3.9.0/spec/tmpl-namespace.yaml @@ -0,0 +1,4 @@ +apiVersion: v1 +kind: Namespace +metadata: + name: ${OPENEBS_NAMESPACE} diff --git a/web/src/installers/versions.js b/web/src/installers/versions.js index 9654a3074c..11d95dfec9 100644 --- a/web/src/installers/versions.js +++ b/web/src/installers/versions.js @@ -493,6 +493,7 @@ module.exports.InstallerVersions = { ], openebs: [ // cron-openebs-update-3 + "3.9.0", "3.8.0", "3.7.0", "3.6.0",