~/src/github.com/openstack-k8s-operators/install_yamls ~/ci-framework-data/artifacts error: the server doesn't have a resource type "openstackversion" bash scripts/gen-namespace.sh + '[' -z /home/zuul/ci-framework-data/artifacts/manifests ']' + '[' -z openstack ']' + OUT_DIR=/home/zuul/ci-framework-data/artifacts/manifests/openstack + '[' '!' -d /home/zuul/ci-framework-data/artifacts/manifests/openstack ']' + cat oc apply -f /home/zuul/ci-framework-data/artifacts/manifests/openstack/namespace.yaml Warning: resource namespaces/openstack is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by oc apply. oc apply should only be used on resources created declaratively by either oc create --save-config or oc apply. The missing annotation will be patched automatically. namespace/openstack configured timeout 500s bash -c "while ! (oc get project.v1.project.openshift.io openstack); do sleep 1; done" NAME DISPLAY NAME STATUS openstack Active oc project openstack Now using project "openstack" on server "https://api.crc.testing:6443". bash scripts/gen-input-kustomize.sh + OUT=/home/zuul/ci-framework-data/artifacts/manifests + '[' -z openstack ']' + '[' -z osp-secret ']' + '[' -z 12345678 ']' + '[' -z 1234567842 ']' + '[' -z 767c3ed056cbaa3b9dfedb8c6f825bf0 ']' + '[' -z sEFmdFjDUqRM2VemYslV5yGNWjokioJXsg8Nrlc3drU= ']' + '[' -z COX8bmlKAWn56XCGMrKQJj7dgHNAOl6f ']' + '[' -z openstack ']' + '[' -z libvirt-secret ']' + DIR=/home/zuul/ci-framework-data/artifacts/manifests/openstack/input + '[' '!' -d /home/zuul/ci-framework-data/artifacts/manifests/openstack/input ']' + mkdir -p /home/zuul/ci-framework-data/artifacts/manifests/openstack/input + pushd /home/zuul/ci-framework-data/artifacts/manifests/openstack/input ~/ci-framework-data/artifacts/manifests/openstack/input ~/src/github.com/openstack-k8s-operators/install_yamls + cat oc get secret/osp-secret || oc kustomize /home/zuul/ci-framework-data/artifacts/manifests/openstack/input | oc apply -f - Error from server (NotFound): secrets "osp-secret" not found secret/libvirt-secret created secret/octavia-ca-passphrase created secret/osp-secret created bash scripts/gen-ceph-kustomize.sh "build" +++ dirname scripts/gen-ceph-kustomize.sh ++ cd scripts ++ pwd -P + SCRIPTPATH=/home/zuul/src/github.com/openstack-k8s-operators/install_yamls/scripts + . /home/zuul/src/github.com/openstack-k8s-operators/install_yamls/scripts/common.sh --source-only ++ set -e + '[' -z quay.io/ceph/demo:latest-squid ']' + '[' -z openstack ']' + '[' '!' -d /home/zuul/ci-framework-data/artifacts/manifests/openstack/ceph/cr ']' + mkdir -p /home/zuul/ci-framework-data/artifacts/manifests/openstack/ceph/cr + pushd /home/zuul/ci-framework-data/artifacts/manifests/openstack/ceph/cr ~/ci-framework-data/artifacts/manifests/openstack/ceph/cr ~/src/github.com/openstack-k8s-operators/install_yamls + CEPH_TIMEOUT=120 + CEPH_HOSTNETWORK=true + CEPH_POOLS=("volumes" "images" "backups" "cephfs.cephfs.meta" "cephfs.cephfs.data") + CEPH_DAEMONS=osd,mds,rgw + CEPH_DATASIZE=10Gi + CEPH_WORKER= + CEPH_MON_CONF= + CEPH_DEMO_UID=0 + OSP_SECRET=osp-secret + RGW_USER=swift + RGW_NAME=ceph ++ oc -n openstack get ingresses.config/cluster -o 'jsonpath={.spec.domain}' + DOMAIN=apps-crc.testing ++ oc -n openstack get secrets osp-secret -o 'jsonpath={.data.SwiftPassword}' ++ base64 -d + RGW_PASS=12345678 + [[ true == \f\a\l\s\e ]] + case "$1" in + bootstrap_ceph ++ oc -n openstack get nodes --selector=node-role.kubernetes.io/worker= -o 'jsonpath={.items[*].status.addresses[*].address}' + NODES='192.168.126.11 crc' + read -ra values + [[ -z '' ]] + MON_IP=192.168.126.11 + '[' -z '' ']' + CEPH_WORKER=crc + add_ceph_pod + cat + ceph_kustomize + cat + '[' -n '' ']' + kustomization_add_resources + echo merge config dir merge config dir ++ find . -type f -name '*.yaml' ++ grep -v kustomization + yamls=./ceph-pod.yaml + for y in ${yamls[@]} + kustomize edit add resource ./ceph-pod.yaml 2025/06/06 09:22:14 resource ./ceph-pod.yaml already in kustomization file bash scripts/operator-deploy-resources.sh + '[' -z /home/zuul/ci-framework-data/artifacts/manifests/openstack/ceph/cr ']' + NEXT_WAIT_TIME=0 + '[' 0 -eq 15 ']' + oc kustomize /home/zuul/ci-framework-data/artifacts/manifests/openstack/ceph/cr + oc apply -f - Warning: would violate PodSecurity "restricted:latest": host namespaces (hostNetwork=true), allowPrivilegeEscalation != false (container "ceph" must set securityContext.allowPrivilegeEscalation=false), unrestricted capabilities (container "ceph" must set securityContext.capabilities.drop=["ALL"]), runAsNonRoot != true (pod or container "ceph" must set securityContext.runAsNonRoot=true), runAsUser=0 (pod must not set runAsUser=0), seccompProfile (pod must not set securityContext.seccompProfile.type to "Unconfined") pod/ceph created + '[' 0 -lt 15 ']' bash scripts/gen-ceph-kustomize.sh "isready" +++ dirname scripts/gen-ceph-kustomize.sh ++ cd scripts ++ pwd -P + SCRIPTPATH=/home/zuul/src/github.com/openstack-k8s-operators/install_yamls/scripts + . /home/zuul/src/github.com/openstack-k8s-operators/install_yamls/scripts/common.sh --source-only ++ set -e + '[' -z quay.io/ceph/demo:latest-squid ']' + '[' -z openstack ']' + '[' '!' -d /home/zuul/ci-framework-data/artifacts/manifests/openstack/ceph/cr ']' + pushd /home/zuul/ci-framework-data/artifacts/manifests/openstack/ceph/cr ~/ci-framework-data/artifacts/manifests/openstack/ceph/cr ~/src/github.com/openstack-k8s-operators/install_yamls + CEPH_TIMEOUT=120 + CEPH_HOSTNETWORK=true + CEPH_POOLS=("volumes" "images" "backups" "cephfs.cephfs.meta" "cephfs.cephfs.data") + CEPH_DAEMONS=osd,mds,rgw + CEPH_DATASIZE=10Gi + CEPH_WORKER= + CEPH_MON_CONF= + CEPH_DEMO_UID=0 + OSP_SECRET=osp-secret + RGW_USER=swift + RGW_NAME=ceph ++ oc -n openstack get ingresses.config/cluster -o 'jsonpath={.spec.domain}' + DOMAIN=apps-crc.testing ++ oc -n openstack get secrets osp-secret -o 'jsonpath={.data.SwiftPassword}' ++ base64 -d + RGW_PASS=12345678 + [[ true == \f\a\l\s\e ]] + case "$1" in + ceph_is_ready + echo 'Waiting the cluster to be up' Waiting the cluster to be up + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 119 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 118 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 117 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 116 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 115 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 114 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 113 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 112 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 111 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 110 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 109 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 108 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 107 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 106 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 105 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 104 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 103 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 102 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 101 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 100 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 99 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 98 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 97 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 96 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 95 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 94 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 93 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 92 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 91 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 90 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 89 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 88 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 87 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 86 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 85 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 84 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 83 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 82 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 81 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 80 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 79 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 78 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 77 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 76 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 75 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 74 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 73 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + sleep 1 + echo -n . .+ (( CEPH_TIMEOUT-- )) + [[ 72 -eq 0 ]] + oc rsh -n openstack ceph ls /etc/ceph/I_AM_A_DEMO + echo bash scripts/gen-ceph-kustomize.sh "config" +++ dirname scripts/gen-ceph-kustomize.sh ++ cd scripts ++ pwd -P + SCRIPTPATH=/home/zuul/src/github.com/openstack-k8s-operators/install_yamls/scripts + . /home/zuul/src/github.com/openstack-k8s-operators/install_yamls/scripts/common.sh --source-only ++ set -e + '[' -z quay.io/ceph/demo:latest-squid ']' + '[' -z openstack ']' + '[' '!' -d /home/zuul/ci-framework-data/artifacts/manifests/openstack/ceph/cr ']' + pushd /home/zuul/ci-framework-data/artifacts/manifests/openstack/ceph/cr ~/ci-framework-data/artifacts/manifests/openstack/ceph/cr ~/src/github.com/openstack-k8s-operators/install_yamls + CEPH_TIMEOUT=120 + CEPH_HOSTNETWORK=true + CEPH_POOLS=("volumes" "images" "backups" "cephfs.cephfs.meta" "cephfs.cephfs.data") + CEPH_DAEMONS=osd,mds,rgw + CEPH_DATASIZE=10Gi + CEPH_WORKER= + CEPH_MON_CONF= + CEPH_DEMO_UID=0 + OSP_SECRET=osp-secret + RGW_USER=swift + RGW_NAME=ceph ++ oc -n openstack get ingresses.config/cluster -o 'jsonpath={.spec.domain}' + DOMAIN=apps-crc.testing ++ oc -n openstack get secrets osp-secret -o 'jsonpath={.data.SwiftPassword}' ++ base64 -d + RGW_PASS=12345678 + [[ true == \f\a\l\s\e ]] + case "$1" in + config_ceph + echo 'Apply Ceph config keys' Apply Ceph config keys + config_keys=(['rgw_keystone_url']='http://keystone-public-openstack.apps-crc.testing' ['rgw_keystone_verify_ssl']='true' ['rgw_keystone_api_version']='3' ['rgw_keystone_accepted_roles']='"member, Member, admin"' ['rgw_keystone_accepted_admin_roles']='"ResellerAdmin, swiftoperator"' ['rgw_keystone_admin_domain']='default' ['rgw_keystone_admin_project']='service' ['rgw_keystone_admin_user']='swift' ['rgw_keystone_admin_password']='12345678' ['rgw_keystone_implicit_tenants']='true' ['rgw_s3_auth_use_keystone']='true' ['rgw_swift_versioning_enabled']='true' ['rgw_swift_enforce_content_length']='true' ['rgw_swift_account_in_url']='true' ['rgw_trust_forwarded_https']='true' ['rgw_max_attr_name_len']='128' ['rgw_max_attrs_num_in_req']='90') + declare -A config_keys + for key in "${!config_keys[@]}" + oc exec -n openstack -it ceph -- sh -c 'ceph config set global rgw_swift_account_in_url true' Unable to use a TTY - input is not a terminal or the right kind of file + for key in "${!config_keys[@]}" + oc exec -n openstack -it ceph -- sh -c 'ceph config set global rgw_keystone_implicit_tenants true' Unable to use a TTY - input is not a terminal or the right kind of file + for key in "${!config_keys[@]}" + oc exec -n openstack -it ceph -- sh -c 'ceph config set global rgw_swift_enforce_content_length true' Unable to use a TTY - input is not a terminal or the right kind of file + for key in "${!config_keys[@]}" + oc exec -n openstack -it ceph -- sh -c 'ceph config set global rgw_keystone_accepted_admin_roles "ResellerAdmin, swiftoperator"' Unable to use a TTY - input is not a terminal or the right kind of file + for key in "${!config_keys[@]}" + oc exec -n openstack -it ceph -- sh -c 'ceph config set global rgw_keystone_admin_domain default' Unable to use a TTY - input is not a terminal or the right kind of file + for key in "${!config_keys[@]}" + oc exec -n openstack -it ceph -- sh -c 'ceph config set global rgw_max_attr_name_len 128' Unable to use a TTY - input is not a terminal or the right kind of file + for key in "${!config_keys[@]}" + oc exec -n openstack -it ceph -- sh -c 'ceph config set global rgw_s3_auth_use_keystone true' Unable to use a TTY - input is not a terminal or the right kind of file + for key in "${!config_keys[@]}" + oc exec -n openstack -it ceph -- sh -c 'ceph config set global rgw_keystone_url http://keystone-public-openstack.apps-crc.testing' Unable to use a TTY - input is not a terminal or the right kind of file + for key in "${!config_keys[@]}" + oc exec -n openstack -it ceph -- sh -c 'ceph config set global rgw_keystone_admin_password 12345678' Unable to use a TTY - input is not a terminal or the right kind of file + for key in "${!config_keys[@]}" + oc exec -n openstack -it ceph -- sh -c 'ceph config set global rgw_keystone_admin_project service' Unable to use a TTY - input is not a terminal or the right kind of file + for key in "${!config_keys[@]}" + oc exec -n openstack -it ceph -- sh -c 'ceph config set global rgw_keystone_accepted_roles "member, Member, admin"' Unable to use a TTY - input is not a terminal or the right kind of file + for key in "${!config_keys[@]}" + oc exec -n openstack -it ceph -- sh -c 'ceph config set global rgw_keystone_admin_user swift' Unable to use a TTY - input is not a terminal or the right kind of file + for key in "${!config_keys[@]}" + oc exec -n openstack -it ceph -- sh -c 'ceph config set global rgw_keystone_verify_ssl true' Unable to use a TTY - input is not a terminal or the right kind of file + for key in "${!config_keys[@]}" + oc exec -n openstack -it ceph -- sh -c 'ceph config set global rgw_max_attrs_num_in_req 90' Unable to use a TTY - input is not a terminal or the right kind of file + for key in "${!config_keys[@]}" + oc exec -n openstack -it ceph -- sh -c 'ceph config set global rgw_trust_forwarded_https true' Unable to use a TTY - input is not a terminal or the right kind of file + for key in "${!config_keys[@]}" + oc exec -n openstack -it ceph -- sh -c 'ceph config set global rgw_keystone_api_version 3' Unable to use a TTY - input is not a terminal or the right kind of file + for key in "${!config_keys[@]}" + oc exec -n openstack -it ceph -- sh -c 'ceph config set global rgw_swift_versioning_enabled true' Unable to use a TTY - input is not a terminal or the right kind of file bash scripts/gen-ceph-kustomize.sh "cephfs" +++ dirname scripts/gen-ceph-kustomize.sh ++ cd scripts ++ pwd -P + SCRIPTPATH=/home/zuul/src/github.com/openstack-k8s-operators/install_yamls/scripts + . /home/zuul/src/github.com/openstack-k8s-operators/install_yamls/scripts/common.sh --source-only ++ set -e + '[' -z quay.io/ceph/demo:latest-squid ']' + '[' -z openstack ']' + '[' '!' -d /home/zuul/ci-framework-data/artifacts/manifests/openstack/ceph/cr ']' + pushd /home/zuul/ci-framework-data/artifacts/manifests/openstack/ceph/cr ~/ci-framework-data/artifacts/manifests/openstack/ceph/cr ~/src/github.com/openstack-k8s-operators/install_yamls + CEPH_TIMEOUT=120 + CEPH_HOSTNETWORK=true + CEPH_POOLS=("volumes" "images" "backups" "cephfs.cephfs.meta" "cephfs.cephfs.data") + CEPH_DAEMONS=osd,mds,rgw + CEPH_DATASIZE=10Gi + CEPH_WORKER= + CEPH_MON_CONF= + CEPH_DEMO_UID=0 + OSP_SECRET=osp-secret + RGW_USER=swift + RGW_NAME=ceph ++ oc -n openstack get ingresses.config/cluster -o 'jsonpath={.spec.domain}' + DOMAIN=apps-crc.testing ++ base64 -d ++ oc -n openstack get secrets osp-secret -o 'jsonpath={.data.SwiftPassword}' + RGW_PASS=12345678 + [[ true == \f\a\l\s\e ]] + case "$1" in + create_volume + echo 'Creating cephfs volume' Creating cephfs volume + oc rsh -n openstack ceph ceph fs volume create cephfs Error EINVAL: filesystem already exists with name 'cephfs' command terminated with exit code 22 + true bash scripts/gen-ceph-kustomize.sh "pools" +++ dirname scripts/gen-ceph-kustomize.sh ++ cd scripts ++ pwd -P + SCRIPTPATH=/home/zuul/src/github.com/openstack-k8s-operators/install_yamls/scripts + . /home/zuul/src/github.com/openstack-k8s-operators/install_yamls/scripts/common.sh --source-only ++ set -e + '[' -z quay.io/ceph/demo:latest-squid ']' + '[' -z openstack ']' + '[' '!' -d /home/zuul/ci-framework-data/artifacts/manifests/openstack/ceph/cr ']' + pushd /home/zuul/ci-framework-data/artifacts/manifests/openstack/ceph/cr ~/ci-framework-data/artifacts/manifests/openstack/ceph/cr ~/src/github.com/openstack-k8s-operators/install_yamls + CEPH_TIMEOUT=120 + CEPH_HOSTNETWORK=true + CEPH_POOLS=("volumes" "images" "backups" "cephfs.cephfs.meta" "cephfs.cephfs.data") + CEPH_DAEMONS=osd,mds,rgw + CEPH_DATASIZE=10Gi + CEPH_WORKER= + CEPH_MON_CONF= + CEPH_DEMO_UID=0 + OSP_SECRET=osp-secret + RGW_USER=swift + RGW_NAME=ceph ++ oc -n openstack get ingresses.config/cluster -o 'jsonpath={.spec.domain}' + DOMAIN=apps-crc.testing ++ oc -n openstack get secrets osp-secret -o 'jsonpath={.data.SwiftPassword}' ++ base64 -d + RGW_PASS=12345678 + [[ true == \f\a\l\s\e ]] + case "$1" in + create_pool + '[' 5 -eq 0 ']' + for pool in "${CEPH_POOLS[@]}" + app=rbd + oc rsh -n openstack ceph ceph osd pool create volumes 4 pool 'volumes' created + [[ volumes = *\c\e\p\h\f\s* ]] + oc rsh -n openstack ceph ceph osd pool application enable volumes rbd enabled application 'rbd' on pool 'volumes' + for pool in "${CEPH_POOLS[@]}" + app=rbd + oc rsh -n openstack ceph ceph osd pool create images 4 pool 'images' created + [[ images = *\c\e\p\h\f\s* ]] + oc rsh -n openstack ceph ceph osd pool application enable images rbd enabled application 'rbd' on pool 'images' + for pool in "${CEPH_POOLS[@]}" + app=rbd + oc rsh -n openstack ceph ceph osd pool create backups 4 pool 'backups' created + [[ backups = *\c\e\p\h\f\s* ]] + oc rsh -n openstack ceph ceph osd pool application enable backups rbd enabled application 'rbd' on pool 'backups' + for pool in "${CEPH_POOLS[@]}" + app=rbd + oc rsh -n openstack ceph ceph osd pool create cephfs.cephfs.meta 4 pool 'cephfs.cephfs.meta' already exists + [[ cephfs.cephfs.meta = *\c\e\p\h\f\s* ]] + app=cephfs + oc rsh -n openstack ceph ceph osd pool application enable cephfs.cephfs.meta cephfs enabled application 'cephfs' on pool 'cephfs.cephfs.meta' + for pool in "${CEPH_POOLS[@]}" + app=rbd + oc rsh -n openstack ceph ceph osd pool create cephfs.cephfs.data 4 pool 'cephfs.cephfs.data' already exists + [[ cephfs.cephfs.data = *\c\e\p\h\f\s* ]] + app=cephfs + oc rsh -n openstack ceph ceph osd pool application enable cephfs.cephfs.data cephfs enabled application 'cephfs' on pool 'cephfs.cephfs.data' bash scripts/gen-ceph-kustomize.sh "secret" +++ dirname scripts/gen-ceph-kustomize.sh ++ cd scripts ++ pwd -P + SCRIPTPATH=/home/zuul/src/github.com/openstack-k8s-operators/install_yamls/scripts + . /home/zuul/src/github.com/openstack-k8s-operators/install_yamls/scripts/common.sh --source-only ++ set -e + '[' -z quay.io/ceph/demo:latest-squid ']' + '[' -z openstack ']' + '[' '!' -d /home/zuul/ci-framework-data/artifacts/manifests/openstack/ceph/cr ']' + pushd /home/zuul/ci-framework-data/artifacts/manifests/openstack/ceph/cr ~/ci-framework-data/artifacts/manifests/openstack/ceph/cr ~/src/github.com/openstack-k8s-operators/install_yamls + CEPH_TIMEOUT=120 + CEPH_HOSTNETWORK=true + CEPH_POOLS=("volumes" "images" "backups" "cephfs.cephfs.meta" "cephfs.cephfs.data") + CEPH_DAEMONS=osd,mds,rgw + CEPH_DATASIZE=10Gi + CEPH_WORKER= + CEPH_MON_CONF= + CEPH_DEMO_UID=0 + OSP_SECRET=osp-secret + RGW_USER=swift + RGW_NAME=ceph ++ oc -n openstack get ingresses.config/cluster -o 'jsonpath={.spec.domain}' + DOMAIN=apps-crc.testing ++ base64 -d ++ oc -n openstack get secrets osp-secret -o 'jsonpath={.data.SwiftPassword}' + RGW_PASS=12345678 + [[ true == \f\a\l\s\e ]] + case "$1" in + create_secret ceph-conf-files + SECRET_NAME=ceph-conf-files ++ mktemp -d + TEMPDIR=/tmp/tmp.bcrstka0Vn + local client=client.openstack + trap 'rm -rf -- "$TEMPDIR"' EXIT + echo 'Copying Ceph config files from the container to /tmp/tmp.bcrstka0Vn' Copying Ceph config files from the container to /tmp/tmp.bcrstka0Vn + oc rsync -n openstack ceph:/etc/ceph/ceph.conf /tmp/tmp.bcrstka0Vn WARNING: cannot use rsync: rsync not available in container ceph.conf + echo 'Create OpenStack keyring' Create OpenStack keyring + create_key client.openstack + local client=client.openstack + local caps + local osd_caps + '[' 5 -eq 0 ']' ++ build_caps ++ local CAPS= ++ for pool in "${CEPH_POOLS[@]}" ++ caps='profile rbd pool=volumes' ++ CAPS+='profile rbd pool=volumes,' ++ for pool in "${CEPH_POOLS[@]}" ++ caps='profile rbd pool=images' ++ CAPS+='profile rbd pool=images,' ++ for pool in "${CEPH_POOLS[@]}" ++ caps='profile rbd pool=backups' ++ CAPS+='profile rbd pool=backups,' ++ for pool in "${CEPH_POOLS[@]}" ++ caps='profile rbd pool=cephfs.cephfs.meta' ++ CAPS+='profile rbd pool=cephfs.cephfs.meta,' ++ for pool in "${CEPH_POOLS[@]}" ++ caps='profile rbd pool=cephfs.cephfs.data' ++ CAPS+='profile rbd pool=cephfs.cephfs.data,' ++ echo 'profile rbd pool=volumes,profile rbd pool=images,profile rbd pool=backups,profile rbd pool=cephfs.cephfs.meta,profile rbd pool=cephfs.cephfs.data' + osd_caps='profile rbd pool=volumes,profile rbd pool=images,profile rbd pool=backups,profile rbd pool=cephfs.cephfs.meta,profile rbd pool=cephfs.cephfs.data' + oc rsh -n openstack ceph ceph auth get-or-create client.openstack mgr 'allow *' mon 'profile rbd' osd 'profile rbd pool=volumes,profile rbd pool=images,profile rbd pool=backups,profile rbd pool=cephfs.cephfs.meta,profile rbd pool=cephfs.cephfs.data' + echo 'Copying OpenStack keyring from the container to /tmp/tmp.bcrstka0Vn' Copying OpenStack keyring from the container to /tmp/tmp.bcrstka0Vn + oc rsh -n openstack ceph ceph auth export client.openstack -o /etc/ceph/ceph.client.openstack.keyring + oc rsync -n openstack ceph:/etc/ceph/ceph.client.openstack.keyring /tmp/tmp.bcrstka0Vn WARNING: cannot use rsync: rsync not available in container ceph.client.openstack.keyring + echo 'Replacing openshift secret ceph-conf-files' Replacing openshift secret ceph-conf-files + oc delete secret ceph-conf-files -n openstack + true + oc create secret generic ceph-conf-files --from-file=/tmp/tmp.bcrstka0Vn/ceph.conf --from-file=/tmp/tmp.bcrstka0Vn/ceph.client.openstack.keyring -n openstack secret/ceph-conf-files created + rm -rf -- /tmp/tmp.bcrstka0Vn bash scripts/gen-ceph-kustomize.sh "post" +++ dirname scripts/gen-ceph-kustomize.sh ++ cd scripts ++ pwd -P + SCRIPTPATH=/home/zuul/src/github.com/openstack-k8s-operators/install_yamls/scripts + . /home/zuul/src/github.com/openstack-k8s-operators/install_yamls/scripts/common.sh --source-only ++ set -e + '[' -z quay.io/ceph/demo:latest-squid ']' + '[' -z openstack ']' + '[' '!' -d /home/zuul/ci-framework-data/artifacts/manifests/openstack/ceph/cr ']' + pushd /home/zuul/ci-framework-data/artifacts/manifests/openstack/ceph/cr ~/ci-framework-data/artifacts/manifests/openstack/ceph/cr ~/src/github.com/openstack-k8s-operators/install_yamls + CEPH_TIMEOUT=120 + CEPH_HOSTNETWORK=true + CEPH_POOLS=("volumes" "images" "backups" "cephfs.cephfs.meta" "cephfs.cephfs.data") + CEPH_DAEMONS=osd,mds,rgw + CEPH_DATASIZE=10Gi + CEPH_WORKER= + CEPH_MON_CONF= + CEPH_DEMO_UID=0 + OSP_SECRET=osp-secret + RGW_USER=swift + RGW_NAME=ceph ++ oc -n openstack get ingresses.config/cluster -o 'jsonpath={.spec.domain}' + DOMAIN=apps-crc.testing ++ base64 -d ++ oc -n openstack get secrets osp-secret -o 'jsonpath={.data.SwiftPassword}' + RGW_PASS=12345678 + [[ true == \f\a\l\s\e ]] + case "$1" in + config_rgw + echo 'Restart RGW and reload the config' Restart RGW and reload the config + oc -n openstack rsh ceph pkill radosgw + name=client.rgw.ceph + path=/var/lib/ceph/radosgw/ceph-rgw.ceph/keyring + options=' --default-log-to-stderr=true --err-to-stderr=true --default-log-to-file=false' + oc -n openstack rsh ceph radosgw --cluster ceph --setuser ceph --setgroup ceph ' --default-log-to-stderr=true --err-to-stderr=true --default-log-to-file=false' -n client.rgw.ceph -k /var/lib/ceph/radosgw/ceph-rgw.ceph/keyring ~/ci-framework-data/artifacts